Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
76,892 | 7,548,573,543 | IssuesEvent | 2018-04-18 11:42:20 | rucio/rucio | https://api.github.com/repos/rucio/rucio | closed | Work on the demo | Testing enhancement patch | Motivation
----------
Demo is empty
Modification
------------
Create RSEs, DIDs...
| 1.0 | Work on the demo - Motivation
----------
Demo is empty
Modification
------------
Create RSEs, DIDs...
| test | work on the demo motivation demo is empty modification create rses dids | 1 |
61,627 | 8,531,489,474 | IssuesEvent | 2018-11-04 12:54:58 | 49thSecurityDivision/lab-documentation | https://api.github.com/repos/49thSecurityDivision/lab-documentation | opened | Snipe-IT submit new assest | documentation lab_minions | **Describe what prompted this request**
The process of how we onboard new items when we buy them.
**Describe the solution you'd like**
Show all steps for adding new items for minions.
**Who is the target user for this?**
Minions
**Will it need to be maintained?**
No, unless snipe-it changes
**Who should have access?**
Lab minions
**Are these docs testable? (not all documentation will be testable)**
No
**Will there be sensitive information in these docs?**
No
**Additional context**
Since Snipe-IT is our asset management system, we need to build a workflow around what happens when we get/purchase new item(s).
| 1.0 | Snipe-IT submit new assest - **Describe what prompted this request**
The process of how we onboard new items when we buy them.
**Describe the solution you'd like**
Show all steps for adding new items for minions.
**Who is the target user for this?**
Minions
**Will it need to be maintained?**
No, unless snipe-it changes
**Who should have access?**
Lab minions
**Are these docs testable? (not all documentation will be testable)**
No
**Will there be sensitive information in these docs?**
No
**Additional context**
Since Snipe-IT is our asset management system, we need to build a workflow around what happens when we get/purchase new item(s).
| non_test | snipe it submit new assest describe what prompted this request the process of how we onboard new items when we buy them describe the solution you d like show all steps for adding new items for minions who is the target user for this minions will it need to be maintained no unless snipe it changes who should have access lab minions are these docs testable not all documentation will be testable no will there be sensitive information in these docs no additional context since snipe it is our asset management system we need to build a workflow around what happens when we get purchase new item s | 0 |
133,857 | 18,358,481,922 | IssuesEvent | 2021-10-08 22:24:36 | GabDug/MaterialIconsLibrary | https://api.github.com/repos/GabDug/MaterialIconsLibrary | opened | CVE-2021-32804 (High) detected in tar-6.1.0.tgz | security vulnerability | ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p>
<p>Path to dependency file: MaterialIconsLibrary/package.json</p>
<p>Path to vulnerable library: MaterialIconsLibrary/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- cli-12.0.0.tgz (Root Library)
- pacote-11.3.2.tgz
- :x: **tar-6.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GabDug/MaterialIconsLibrary/commit/89aa5982db5f213f61df052ca5dc7f0732ce3a91">89aa5982db5f213f61df052ca5dc7f0732ce3a91</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-32804 (High) detected in tar-6.1.0.tgz - ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p>
<p>Path to dependency file: MaterialIconsLibrary/package.json</p>
<p>Path to vulnerable library: MaterialIconsLibrary/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- cli-12.0.0.tgz (Root Library)
- pacote-11.3.2.tgz
- :x: **tar-6.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GabDug/MaterialIconsLibrary/commit/89aa5982db5f213f61df052ca5dc7f0732ce3a91">89aa5982db5f213f61df052ca5dc7f0732ce3a91</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file materialiconslibrary package json path to vulnerable library materialiconslibrary node modules tar package json dependency hierarchy cli tgz root library pacote tgz x tar tgz vulnerable library found in head commit a href found in base branch main vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource | 0 |
134,175 | 10,885,485,490 | IssuesEvent | 2019-11-18 10:29:33 | SatelliteQE/robottelo | https://api.github.com/repos/SatelliteQE/robottelo | closed | cli.test_user.UserWithCleanUpTestCase.test_positive_update_all_roles not isolated | 6.3 6.4 Bug test-failure | besides that i don't really know what this test is supposed to really test (it look quite useless to me),
it also lacks the isolation:
It first lists all the roles and then creates an user and assigns the roles to it one by one.
- However, some other role-related test runs in parralel, creating and managing its own custom role, including deleting it at the end.
This causes the first test to discover this role during initial listing and sometimes assigning, however it fails to find it at the end during the assertion listing, since it no longer exists | 1.0 | cli.test_user.UserWithCleanUpTestCase.test_positive_update_all_roles not isolated - besides that i don't really know what this test is supposed to really test (it look quite useless to me),
it also lacks the isolation:
It first lists all the roles and then creates an user and assigns the roles to it one by one.
- However, some other role-related test runs in parralel, creating and managing its own custom role, including deleting it at the end.
This causes the first test to discover this role during initial listing and sometimes assigning, however it fails to find it at the end during the assertion listing, since it no longer exists | test | cli test user userwithcleanuptestcase test positive update all roles not isolated besides that i don t really know what this test is supposed to really test it look quite useless to me it also lacks the isolation it first lists all the roles and then creates an user and assigns the roles to it one by one however some other role related test runs in parralel creating and managing its own custom role including deleting it at the end this causes the first test to discover this role during initial listing and sometimes assigning however it fails to find it at the end during the assertion listing since it no longer exists | 1 |
8,916 | 3,010,872,727 | IssuesEvent | 2015-07-28 15:13:52 | brharp/hjckrrh | https://api.github.com/repos/brharp/hjckrrh | closed | Bootstrap Anchor issue | priority: high status: fixed - testing required type: bug | Platform configuration change needed - When in Home |Administration |Appearance | Settings
in Anchors sections - need to to deselect the "fix anchor positions". Otherwise anchors don't work in bootstrap. | 1.0 | Bootstrap Anchor issue - Platform configuration change needed - When in Home |Administration |Appearance | Settings
in Anchors sections - need to to deselect the "fix anchor positions". Otherwise anchors don't work in bootstrap. | test | bootstrap anchor issue platform configuration change needed when in home administration appearance settings in anchors sections need to to deselect the fix anchor positions otherwise anchors don t work in bootstrap | 1 |
164,382 | 12,801,054,229 | IssuesEvent | 2020-07-02 18:20:37 | syscoin/syscoin | https://api.github.com/repos/syscoin/syscoin | opened | New masternode commands causes crash when entered without argument | 4.2-testnet | Issue most likely due to ported commands being non-standard to the bitcoin's new rpc format
**Expected behavior**
```
$ syscoin-cli bls
error code: -1
error message:
bls
Set of commands to execute BLS related actions.
To get help on individual commands, use "help bls command".
Arguments:
1. "command" (string, required) The command to execute
Available commands:
generate - Create a BLS secret/public key pair
fromsecret - Parse a BLS secret key and return the secret/public key pair
Result:
null (json null)
```
```
$ syscoin-cli help protx
error code: -1
error message:
protx
Set of commands to execute ProTx related actions.
To get help on individual commands, use "help protx command".
Arguments:
1. "command" (string, required) The command to execute
Available commands:
register - Create and send ProTx to network
register_fund - Fund, create and send ProTx to network
register_prepare - Create an unsigned ProTx
register_submit - Sign and submit a ProTx
list - List ProTxs
info - Return information about a ProTx
update_service - Create and send ProUpServTx to network
update_registrar - Create and send ProUpRegTx to network
revoke - Create and send ProUpRevTx to network
diff - Calculate a diff and a proof between two masternode lists
Result:
null (json null)
```
<!--- What behavior did you expect? -->
**Actual behavior**
```
$ syscoin-cli bls
error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached")
Make sure the syscoind server is running and that you are connecting to the correct RPC port.
```
```
$ syscoin-cli protx
error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached")
Make sure the syscoind server is running and that you are connecting to the correct RPC port.
```
<!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? -->
**To reproduce**
`syscoin-cli bls`
`syscoin-cli masternode`
`syscoin-cli protx`
| 1.0 | New masternode commands causes crash when entered without argument - Issue most likely due to ported commands being non-standard to the bitcoin's new rpc format
**Expected behavior**
```
$ syscoin-cli bls
error code: -1
error message:
bls
Set of commands to execute BLS related actions.
To get help on individual commands, use "help bls command".
Arguments:
1. "command" (string, required) The command to execute
Available commands:
generate - Create a BLS secret/public key pair
fromsecret - Parse a BLS secret key and return the secret/public key pair
Result:
null (json null)
```
```
$ syscoin-cli help protx
error code: -1
error message:
protx
Set of commands to execute ProTx related actions.
To get help on individual commands, use "help protx command".
Arguments:
1. "command" (string, required) The command to execute
Available commands:
register - Create and send ProTx to network
register_fund - Fund, create and send ProTx to network
register_prepare - Create an unsigned ProTx
register_submit - Sign and submit a ProTx
list - List ProTxs
info - Return information about a ProTx
update_service - Create and send ProUpServTx to network
update_registrar - Create and send ProUpRegTx to network
revoke - Create and send ProUpRevTx to network
diff - Calculate a diff and a proof between two masternode lists
Result:
null (json null)
```
<!--- What behavior did you expect? -->
**Actual behavior**
```
$ syscoin-cli bls
error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached")
Make sure the syscoind server is running and that you are connecting to the correct RPC port.
```
```
$ syscoin-cli protx
error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached")
Make sure the syscoind server is running and that you are connecting to the correct RPC port.
```
<!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? -->
**To reproduce**
`syscoin-cli bls`
`syscoin-cli masternode`
`syscoin-cli protx`
| test | new masternode commands causes crash when entered without argument issue most likely due to ported commands being non standard to the bitcoin s new rpc format expected behavior syscoin cli bls error code error message bls set of commands to execute bls related actions to get help on individual commands use help bls command arguments command string required the command to execute available commands generate create a bls secret public key pair fromsecret parse a bls secret key and return the secret public key pair result null json null syscoin cli help protx error code error message protx set of commands to execute protx related actions to get help on individual commands use help protx command arguments command string required the command to execute available commands register create and send protx to network register fund fund create and send protx to network register prepare create an unsigned protx register submit sign and submit a protx list list protxs info return information about a protx update service create and send proupservtx to network update registrar create and send proupregtx to network revoke create and send prouprevtx to network diff calculate a diff and a proof between two masternode lists result null json null actual behavior syscoin cli bls error could not connect to the server error code eof reached make sure the syscoind server is running and that you are connecting to the correct rpc port syscoin cli protx error could not connect to the server error code eof reached make sure the syscoind server is running and that you are connecting to the correct rpc port to reproduce syscoin cli bls syscoin cli masternode syscoin cli protx | 1 |
188,063 | 6,767,976,918 | IssuesEvent | 2017-10-26 06:57:47 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | OTP SMS delivery metrics | epic/sms kind/user_story priority/medium status/wontfix | We should have a metrics for SMS delivery process
* Succesful/unsuccessful SMS submissions stats
* Undelivered SMS
* SMS delivery latency
- [ ] integration with life report to store counters
- [ ] new metrics on datadog
https://docs.google.com/spreadsheets/d/1X1gQEWQc02loG1OtNRZzzuN3NssLRoESIgpn-aRDMPQ/edit?usp=sharing | 1.0 | OTP SMS delivery metrics - We should have a metrics for SMS delivery process
* Succesful/unsuccessful SMS submissions stats
* Undelivered SMS
* SMS delivery latency
- [ ] integration with life report to store counters
- [ ] new metrics on datadog
https://docs.google.com/spreadsheets/d/1X1gQEWQc02loG1OtNRZzzuN3NssLRoESIgpn-aRDMPQ/edit?usp=sharing | non_test | otp sms delivery metrics we should have a metrics for sms delivery process succesful unsuccessful sms submissions stats undelivered sms sms delivery latency integration with life report to store counters new metrics on datadog | 0 |
121,715 | 10,193,878,281 | IssuesEvent | 2019-08-12 14:34:01 | sparkdesignsystem/spark-design-system | https://api.github.com/repos/sparkdesignsystem/spark-design-system | closed | Follow/successfully pass Manual Test Plan for Promo | scope: all pkgs status: PO approved type: testing | **User Story:**
As Spark we want to follow the Manual Test plan for Promo so that we can confirm Manual Test plan is up to date.
**AC:**
- Successful complete Manual Test Plan for Promo | 1.0 | Follow/successfully pass Manual Test Plan for Promo - **User Story:**
As Spark we want to follow the Manual Test plan for Promo so that we can confirm Manual Test plan is up to date.
**AC:**
- Successful complete Manual Test Plan for Promo | test | follow successfully pass manual test plan for promo user story as spark we want to follow the manual test plan for promo so that we can confirm manual test plan is up to date ac successful complete manual test plan for promo | 1 |
6,655 | 3,039,344,505 | IssuesEvent | 2015-08-07 08:13:40 | spring-projects/spring-boot | https://api.github.com/repos/spring-projects/spring-boot | closed | reference documentation "67.7 Use Two EntityManagers" should be extended | documentation | the reference documentation "67.7 Use Two EntityManagers" describe the declaration for a second EntityManagerFactory, but the usage is obmitted. usage examples should be added.
for example with spring data jpa:
```
@Configuration
@EnableJpaRepositories(value = "...dataaccess.customer.repository",
entityManagerFactoryRef = "customerEntityManagerFactory")
public class CustomerInitializer {}
```
```
@Configuration
@EnableJpaRepositories(value = "...dataaccess.order.repository",
entityManagerFactoryRef = "orderEntityManagerFactory")
public class OrderInitializer {}
```
thx Rico | 1.0 | reference documentation "67.7 Use Two EntityManagers" should be extended - the reference documentation "67.7 Use Two EntityManagers" describe the declaration for a second EntityManagerFactory, but the usage is obmitted. usage examples should be added.
for example with spring data jpa:
```
@Configuration
@EnableJpaRepositories(value = "...dataaccess.customer.repository",
entityManagerFactoryRef = "customerEntityManagerFactory")
public class CustomerInitializer {}
```
```
@Configuration
@EnableJpaRepositories(value = "...dataaccess.order.repository",
entityManagerFactoryRef = "orderEntityManagerFactory")
public class OrderInitializer {}
```
thx Rico | non_test | reference documentation use two entitymanagers should be extended the reference documentation use two entitymanagers describe the declaration for a second entitymanagerfactory but the usage is obmitted usage examples should be added for example with spring data jpa configuration enablejparepositories value dataaccess customer repository entitymanagerfactoryref customerentitymanagerfactory public class customerinitializer configuration enablejparepositories value dataaccess order repository entitymanagerfactoryref orderentitymanagerfactory public class orderinitializer thx rico | 0 |
185,396 | 14,351,682,038 | IssuesEvent | 2020-11-30 02:03:54 | CSU-Booking-Platform/application | https://api.github.com/repos/CSU-Booking-Platform/application | opened | Acceptance Tests for #34 Create Permissions | acceptance-test | ### User story
#34
### Acceptance Test Checklist
- [ ] If I am an administrator, when I try to create a permission, the system registers a new permission
- [ ] If I am not an administrator, when I try to create a permission, the system does not create the new permission
| 1.0 | Acceptance Tests for #34 Create Permissions - ### User story
#34
### Acceptance Test Checklist
- [ ] If I am an administrator, when I try to create a permission, the system registers a new permission
- [ ] If I am not an administrator, when I try to create a permission, the system does not create the new permission
| test | acceptance tests for create permissions user story acceptance test checklist if i am an administrator when i try to create a permission the system registers a new permission if i am not an administrator when i try to create a permission the system does not create the new permission | 1 |
310,022 | 26,694,645,965 | IssuesEvent | 2023-01-27 09:17:31 | MetaMask/metamask-mobile | https://api.github.com/repos/MetaMask/metamask-mobile | opened | Improve test coverage of send flow | tests team-confirmations | Improve test coverage of send flow on mobile. This includes:
1. Writing more unit test coverage for different components used in send flow
2. Writing more e2e test for send screens:
 | 1.0 | Improve test coverage of send flow - Improve test coverage of send flow on mobile. This includes:
1. Writing more unit test coverage for different components used in send flow
2. Writing more e2e test for send screens:
 | test | improve test coverage of send flow improve test coverage of send flow on mobile this includes writing more unit test coverage for different components used in send flow writing more test for send screens | 1 |
71,438 | 7,244,815,618 | IssuesEvent | 2018-02-14 16:11:02 | eclipse/kura | https://api.github.com/repos/eclipse/kura | closed | [Testing] BLE Util test erratic | testing | BluetoothLeUtilTest.testKillCmd has a tendency to fail on CI. Remove it. | 1.0 | [Testing] BLE Util test erratic - BluetoothLeUtilTest.testKillCmd has a tendency to fail on CI. Remove it. | test | ble util test erratic bluetoothleutiltest testkillcmd has a tendency to fail on ci remove it | 1 |
161,134 | 13,806,692,518 | IssuesEvent | 2020-10-11 18:44:19 | codimd/server | https://api.github.com/repos/codimd/server | closed | Add note revisions documentation | documentation enhancement good first issue | After getting #195 and checking out features page, I had to notice, that we don't explain note revisions anywhere.

We should add some documentation about them to the [features.md](https://github.com/codimd/server/blob/master/public/docs/features.md). | 1.0 | Add note revisions documentation - After getting #195 and checking out features page, I had to notice, that we don't explain note revisions anywhere.

We should add some documentation about them to the [features.md](https://github.com/codimd/server/blob/master/public/docs/features.md). | non_test | add note revisions documentation after getting and checking out features page i had to notice that we don t explain note revisions anywhere we should add some documentation about them to the | 0 |
226,055 | 17,937,153,821 | IssuesEvent | 2021-09-10 16:48:48 | IntellectualSites/FastAsyncWorldEdit | https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit | closed | //regen Spitting Errors in Console when attempting to execute command | Requires Testing | ### Server Implementation
Spigot
### Server Version
1.17.1
### Describe the bug
When attempting to use //regen after selecting an area, the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console.
Some of the effects of the error are seen ingame, such as grass not generating on the surface or bedrock not generating. Cave generation will also be slightly different.
Here is a paste of the error in console: [](https://pastebin.com/gyysY9Ft)
### To Reproduce
1. Select an area larger than 2000 blocks. Try to include the surface grass or bedrock for the clearest effect.
2. Run //regen ingame.
3. Errors should appear in console and the regenerated region will look different than the original vanilla terrain.
### Expected behaviour
After using //regen I was expecting the area to generate as if it were being generated by minecraft for the first time, with identical structure and landform.
### Screenshots / Videos


### Error log (if applicable)
https://pastebin.com/gyysY9Ft
### Fawe Debugpaste
https://athion.net/ISPaster/paste/view/b12a1d36b6b544ffb7890ad1fa2b091e
### Fawe Version
FastAsyncWorldEdit version 1.17-217;c16212e
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit-1.17/ and the issue still persists.
### Anything else?
I've tested //regen in multiple areas of the overworld, and the issue remains consistent. I also am running an anticheat on the server (Vulcan), but I highly doubt this is the cause of the issue. | 1.0 | //regen Spitting Errors in Console when attempting to execute command - ### Server Implementation
Spigot
### Server Version
1.17.1
### Describe the bug
When attempting to use //regen after selecting an area, the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console.
Some of the effects of the error are seen ingame, such as grass not generating on the surface or bedrock not generating. Cave generation will also be slightly different.
Here is a paste of the error in console: [](https://pastebin.com/gyysY9Ft)
### To Reproduce
1. Select an area larger than 2000 blocks. Try to include the surface grass or bedrock for the clearest effect.
2. Run //regen ingame.
3. Errors should appear in console and the regenerated region will look different than the original vanilla terrain.
### Expected behaviour
After using //regen I was expecting the area to generate as if it were being generated by minecraft for the first time, with identical structure and landform.
### Screenshots / Videos


### Error log (if applicable)
https://pastebin.com/gyysY9Ft
### Fawe Debugpaste
https://athion.net/ISPaster/paste/view/b12a1d36b6b544ffb7890ad1fa2b091e
### Fawe Version
FastAsyncWorldEdit version 1.17-217;c16212e
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit-1.17/ and the issue still persists.
### Anything else?
I've tested //regen in multiple areas of the overworld, and the issue remains consistent. I also am running an anticheat on the server (Vulcan), but I highly doubt this is the cause of the issue. | test | regen spitting errors in console when attempting to execute command server implementation spigot server version describe the bug when attempting to use regen after selecting an area the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console some of the effects of the error are seen ingame such as grass not generating on the surface or bedrock not generating cave generation will also be slightly different here is a paste of the error in console to reproduce select an area larger than blocks try to include the surface grass or bedrock for the clearest effect run regen ingame errors should appear in console and the regenerated region will look different than the original vanilla terrain expected behaviour after using regen i was expecting the area to generate as if it were being generated by minecraft for the first time with identical structure and landform screenshots videos error log if applicable fawe debugpaste fawe version fastasyncworldedit version checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else i ve tested regen in multiple areas of the overworld and the issue remains consistent i also am running an anticheat on the server vulcan but i highly doubt this is the cause of the issue | 1 |
234,833 | 19,270,882,863 | IssuesEvent | 2021-12-10 05:09:06 | cse110-fa21-group2/cse110-fa21-group2 | https://api.github.com/repos/cse110-fa21-group2/cse110-fa21-group2 | closed | Laurence's Testing Assignments | testing | #Functions to Test
- function clearSortingandFiltering
- async function populateExplore()
- function initializeButtons()
- createCardsFromData(items, container)
- function initializeLocalStorage() | 1.0 | Laurence's Testing Assignments - #Functions to Test
- function clearSortingandFiltering
- async function populateExplore()
- function initializeButtons()
- createCardsFromData(items, container)
- function initializeLocalStorage() | test | laurence s testing assignments functions to test function clearsortingandfiltering async function populateexplore function initializebuttons createcardsfromdata items container function initializelocalstorage | 1 |
265,177 | 23,149,609,072 | IssuesEvent | 2022-07-29 06:54:07 | cnosdb/cnosdb | https://api.github.com/repos/cnosdb/cnosdb | closed | [Bug]parallel kvcore test will panic | type/bug test kv store | **Describe the bug**
parallel kvcore test will panic,we add serial for each test,the test will pass
**To Reproduce**
cargo test --package tskv --lib kvcore::test
**Expected behavior**
should parallel
**Additional context**
none | 1.0 | [Bug]parallel kvcore test will panic - **Describe the bug**
parallel kvcore test will panic,we add serial for each test,the test will pass
**To Reproduce**
cargo test --package tskv --lib kvcore::test
**Expected behavior**
should parallel
**Additional context**
none | test | parallel kvcore test will panic describe the bug parallel kvcore test will panic,we add serial for each test,the test will pass to reproduce cargo test package tskv lib kvcore test expected behavior should parallel additional context none | 1 |
148,000 | 19,526,266,752 | IssuesEvent | 2021-12-30 08:25:55 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2020-25672 (High) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2020-25672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak vulnerability was found in Linux kernel in llcp_sock_connect
<p>Publish Date: 2021-05-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-25672 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2020-25672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak vulnerability was found in Linux kernel in llcp_sock_connect
<p>Publish Date: 2021-05-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net nfc llcp sock c net nfc llcp sock c vulnerability details a memory leak vulnerability was found in linux kernel in llcp sock connect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
244,967 | 20,735,263,055 | IssuesEvent | 2022-03-14 13:12:24 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | Increase Unit Test Coverage for Azure Storage Blob | Storage Client help wanted test-enhancement | Add additional unit tests to get closer to 90% coverage.
https://azuresdkartifacts.blob.core.windows.net/azure-sdk-for-java/test-coverage/azure-storage-blob/index.html#dn-c | 1.0 | Increase Unit Test Coverage for Azure Storage Blob - Add additional unit tests to get closer to 90% coverage.
https://azuresdkartifacts.blob.core.windows.net/azure-sdk-for-java/test-coverage/azure-storage-blob/index.html#dn-c | test | increase unit test coverage for azure storage blob add additional unit tests to get closer to coverage | 1 |
285,438 | 24,666,612,473 | IssuesEvent | 2022-10-18 10:46:24 | strimzi/strimzi-kafka-operator | https://api.github.com/repos/strimzi/strimzi-kafka-operator | closed | [ST] Add message count, kafka + zk + eo STS or deployment names into the `TestStorage` | enhancement System tests | **Current situation**
Currently we are using `TestStorage` in most of cases, but this storage doesn't contains information about message count (for clients exchange), Kafka & ZK & EO `StatefulSet` or `Deployment` names. It would be much better to add those (and others, which are needed) to the `TestStorage`, so we would be able to remove all extra variables from the tests and use just the `TestStorage`.
**Describe the solution you'd like**
Best solution would be to add, how I mentioned, the all needed information to the `TestStorage` and customize the methods with it - that means, that for example `ClientUtils.waitForClientSuccess` could work just with `TestStorage` and not the multiple other parameters - which is sometimes confusing and it could cause some future issues. | 1.0 | [ST] Add message count, kafka + zk + eo STS or deployment names into the `TestStorage` - **Current situation**
Currently we are using `TestStorage` in most of cases, but this storage doesn't contains information about message count (for clients exchange), Kafka & ZK & EO `StatefulSet` or `Deployment` names. It would be much better to add those (and others, which are needed) to the `TestStorage`, so we would be able to remove all extra variables from the tests and use just the `TestStorage`.
**Describe the solution you'd like**
Best solution would be to add, how I mentioned, the all needed information to the `TestStorage` and customize the methods with it - that means, that for example `ClientUtils.waitForClientSuccess` could work just with `TestStorage` and not the multiple other parameters - which is sometimes confusing and it could cause some future issues. | test | add message count kafka zk eo sts or deployment names into the teststorage current situation currently we are using teststorage in most of cases but this storage doesn t contains information about message count for clients exchange kafka zk eo statefulset or deployment names it would be much better to add those and others which are needed to the teststorage so we would be able to remove all extra variables from the tests and use just the teststorage describe the solution you d like best solution would be to add how i mentioned the all needed information to the teststorage and customize the methods with it that means that for example clientutils waitforclientsuccess could work just with teststorage and not the multiple other parameters which is sometimes confusing and it could cause some future issues | 1 |
152,364 | 12,102,649,016 | IssuesEvent | 2020-04-20 17:02:34 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | cluster k8s >=1.15 with Canal/Calico do not deploy in 2.3.0 and 2.3.2 | [zube]: To Test kind/bug-qa | **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Deploy cluster using k8s - 1.15, 1.16 and 1.17
- The cluster fails to come up
- etcd container logs:
```
2020-03-28 01:44:40.270970 I | embed: rejected connection from "ip:42920" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:46:53.878418 I | embed: rejected connection from "ip:41891" (error "EOF", ServerName "")
2020-03-28 01:47:05.883684 I | embed: rejected connection from "ip:42954" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:47:11.027567 I | embed: rejected connection from "ip:42958" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:47:16.171899 I | embed: rejected connection from "ip:42960" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
```
**Expected Result:**
cluster should be deployed successfully
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.3.0 to 2.3.2
- Installation option (single install/HA): single
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do
- Kubernetes version (use `kubectl version`):
```
1.15, 1.16 and 1.17
```
| 1.0 | cluster k8s >=1.15 with Canal/Calico do not deploy in 2.3.0 and 2.3.2 - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Deploy cluster using k8s - 1.15, 1.16 and 1.17
- The cluster fails to come up
- etcd container logs:
```
2020-03-28 01:44:40.270970 I | embed: rejected connection from "ip:42920" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:46:53.878418 I | embed: rejected connection from "ip:41891" (error "EOF", ServerName "")
2020-03-28 01:47:05.883684 I | embed: rejected connection from "ip:42954" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:47:11.027567 I | embed: rejected connection from "ip:42958" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
2020-03-28 01:47:16.171899 I | embed: rejected connection from "ip:42960" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "")
```
**Expected Result:**
cluster should be deployed successfully
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.3.0 to 2.3.2
- Installation option (single install/HA): single
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do
- Kubernetes version (use `kubectl version`):
```
1.15, 1.16 and 1.17
```
| test | cluster with canal calico do not deploy in and what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible deploy cluster using and the cluster fails to come up etcd container logs i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error eof servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername expected result cluster should be deployed successfully environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui to installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported rke do kubernetes version use kubectl version and | 1 |
318,805 | 27,321,752,661 | IssuesEvent | 2023-02-24 20:32:08 | celestiaorg/celestia-app | https://api.github.com/repos/celestiaorg/celestia-app | closed | `nmt_wrapper_test.go` depends on magic constant | testing | ## Context
https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper_test.go#L98
## Problem
Based on [these lines](https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper.go#L75-L80) I expect the entire share (i.e. 512 bytes) to be pushed to the erasured NMT. `appconsts.ContinuationSparseShareContentSize` + 1 may not always == the share size - the len of the namespace | 1.0 | `nmt_wrapper_test.go` depends on magic constant - ## Context
https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper_test.go#L98
## Problem
Based on [these lines](https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper.go#L75-L80) I expect the entire share (i.e. 512 bytes) to be pushed to the erasured NMT. `appconsts.ContinuationSparseShareContentSize` + 1 may not always == the share size - the len of the namespace | test | nmt wrapper test go depends on magic constant context problem based on i expect the entire share i e bytes to be pushed to the erasured nmt appconsts continuationsparsesharecontentsize may not always the share size the len of the namespace | 1 |
229,713 | 25,362,329,292 | IssuesEvent | 2022-11-21 01:08:08 | Killy85/game_ai_trainer | https://api.github.com/repos/Killy85/game_ai_trainer | opened | CVE-2022-41896 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2022-41896 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. If `ThreadUnsafeUnigramCandidateSampler` is given input `filterbank_channel_count` greater than the allowed max size, TensorFlow will crash. We have patched the issue in GitHub commit 39ec7eaf1428e90c37787e5b3fbd68ebd3c48860. The fix will be included in TensorFlow 2.11. We will also cherrypick this commit on TensorFlow 2.10.1, 2.9.3, and TensorFlow 2.8.4, as these are also affected and still in supported range.
<p>Publish Date: 2022-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41896>CVE-2022-41896</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-41896">https://www.cve.org/CVERecord?id=CVE-2022-41896</a></p>
<p>Release Date: 2022-11-18</p>
<p>Fix Resolution: tensorflow - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-cpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-gpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-41896 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2022-41896 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. If `ThreadUnsafeUnigramCandidateSampler` is given input `filterbank_channel_count` greater than the allowed max size, TensorFlow will crash. We have patched the issue in GitHub commit 39ec7eaf1428e90c37787e5b3fbd68ebd3c48860. The fix will be included in TensorFlow 2.11. We will also cherrypick this commit on TensorFlow 2.10.1, 2.9.3, and TensorFlow 2.8.4, as these are also affected and still in supported range.
<p>Publish Date: 2022-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41896>CVE-2022-41896</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-41896">https://www.cve.org/CVERecord?id=CVE-2022-41896</a></p>
<p>Release Date: 2022-11-18</p>
<p>Fix Resolution: tensorflow - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-cpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-gpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning if threadunsafeunigramcandidatesampler is given input filterbank channel count greater than the allowed max size tensorflow will crash we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend | 0 |
294,786 | 25,404,618,572 | IssuesEvent | 2022-11-22 14:29:45 | eclipse-openj9/openj9 | https://api.github.com/repos/eclipse-openj9/openj9 | opened | SC_Softmx_JitAot_Linux hang | comp:jit test failure | https://openj9-jenkins.osuosl.org/job/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324
SC_Softmx_JitAot_Linux_1 `-Xcompressedrefs -Xjit -Xgcpolicy:gencon `
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324/system_test_output.tar.gz
```
STF 01:15:28.337 - **FAILED** Process jvm4 has timed out
STF 01:15:28.337 - Collecting dumps for: jvm4
```
There are core/javacore/etc. diagnostic files. The javacore files show this state:
```
2LKREGMON JIT-QueueSlotMonitor-1 lock (0x0000FFFF94003300): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "load-1" (J9VMThread:0x000000000057C200)
3LKWAITNOTIFY "load-2" (J9VMThread:0x000000000057E700)
3LKWAITNOTIFY "load-3" (J9VMThread:0x0000000000580B00)
3LKWAITNOTIFY "load-4" (J9VMThread:0x0000000000583000)
3LKWAITNOTIFY "load-5" (J9VMThread:0x0000000000585500)
3LKWAITNOTIFY "load-6" (J9VMThread:0x0000000000587A00)
2LKREGMON JIT-QueueSlotMonitor-4 lock (0x0000FFFF60002610): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "main" (J9VMThread:0x0000000000011200)
2LKREGMON JIT-QueueSlotMonitor-5 lock (0x0000FFFF600026C8): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "load-0" (J9VMThread:0x00000000003E6E00)
``` | 1.0 | SC_Softmx_JitAot_Linux hang - https://openj9-jenkins.osuosl.org/job/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324
SC_Softmx_JitAot_Linux_1 `-Xcompressedrefs -Xjit -Xgcpolicy:gencon `
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324/system_test_output.tar.gz
```
STF 01:15:28.337 - **FAILED** Process jvm4 has timed out
STF 01:15:28.337 - Collecting dumps for: jvm4
```
There are core/javacore/etc. diagnostic files. The javacore files show this state:
```
2LKREGMON JIT-QueueSlotMonitor-1 lock (0x0000FFFF94003300): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "load-1" (J9VMThread:0x000000000057C200)
3LKWAITNOTIFY "load-2" (J9VMThread:0x000000000057E700)
3LKWAITNOTIFY "load-3" (J9VMThread:0x0000000000580B00)
3LKWAITNOTIFY "load-4" (J9VMThread:0x0000000000583000)
3LKWAITNOTIFY "load-5" (J9VMThread:0x0000000000585500)
3LKWAITNOTIFY "load-6" (J9VMThread:0x0000000000587A00)
2LKREGMON JIT-QueueSlotMonitor-4 lock (0x0000FFFF60002610): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "main" (J9VMThread:0x0000000000011200)
2LKREGMON JIT-QueueSlotMonitor-5 lock (0x0000FFFF600026C8): <unowned>
3LKNOTIFYQ Waiting to be notified:
3LKWAITNOTIFY "load-0" (J9VMThread:0x00000000003E6E00)
``` | test | sc softmx jitaot linux hang sc softmx jitaot linux xcompressedrefs xjit xgcpolicy gencon stf failed process has timed out stf collecting dumps for there are core javacore etc diagnostic files the javacore files show this state jit queueslotmonitor lock waiting to be notified load load load load load load jit queueslotmonitor lock waiting to be notified main jit queueslotmonitor lock waiting to be notified load | 1 |
63,429 | 6,846,934,407 | IssuesEvent | 2017-11-13 13:54:41 | Shadowss/TravianZ | https://api.github.com/repos/Shadowss/TravianZ | closed | Warning with language display | bug needs testing | ```
Warning: include_once(GameEngine/lang/en.php): failed to open stream: No such file or directory in /home/u997437127/public_html/GameEngine/Technology.php on line 33
Warning: include_once(): Failed opening 'GameEngine/lang/en.php' for inclusion (include_path='.:/opt/php-7.0/pear') in /home/u997437127/public_html/GameEngine/Technology.php on line 33
```
Need to replace `include_once($autoprefix."GameEngine/lang/en.php");`
Should be `include_once($autoprefix."GameEngine/Lang/en.php");`
https://github.com/Shadowss/TravianZ/blob/994cef151cf9da8b8df8d7f7f58aa60b1dc2e143/GameEngine/Technology.php#L33 | 1.0 | Warning with language display - ```
Warning: include_once(GameEngine/lang/en.php): failed to open stream: No such file or directory in /home/u997437127/public_html/GameEngine/Technology.php on line 33
Warning: include_once(): Failed opening 'GameEngine/lang/en.php' for inclusion (include_path='.:/opt/php-7.0/pear') in /home/u997437127/public_html/GameEngine/Technology.php on line 33
```
Need to replace `include_once($autoprefix."GameEngine/lang/en.php");`
Should be `include_once($autoprefix."GameEngine/Lang/en.php");`
https://github.com/Shadowss/TravianZ/blob/994cef151cf9da8b8df8d7f7f58aa60b1dc2e143/GameEngine/Technology.php#L33 | test | warning with language display warning include once gameengine lang en php failed to open stream no such file or directory in home public html gameengine technology php on line warning include once failed opening gameengine lang en php for inclusion include path opt php pear in home public html gameengine technology php on line need to replace include once autoprefix gameengine lang en php should be include once autoprefix gameengine lang en php | 1 |
352,563 | 32,076,821,348 | IssuesEvent | 2023-09-25 11:34:58 | taskany-inc/issues | https://api.github.com/repos/taskany-inc/issues | closed | E2E: Project / | tests need info | - title, parent, description exists and correct
- goal tab is active
- settings tab is visible for owner and admin only
- add goal button exists and opens new goal modal (currently is not — must be fixed) | 1.0 | E2E: Project / - - title, parent, description exists and correct
- goal tab is active
- settings tab is visible for owner and admin only
- add goal button exists and opens new goal modal (currently is not — must be fixed) | test | project title parent description exists and correct goal tab is active settings tab is visible for owner and admin only add goal button exists and opens new goal modal currently is not — must be fixed | 1 |
266,961 | 20,173,694,233 | IssuesEvent | 2022-02-10 12:44:43 | halkyonio/tap | https://api.github.com/repos/halkyonio/tap | opened | Ad dthe steps to bind the quarkus app to a service | documentation enhancement | ## TODO
Add the steps to bind the Quarkus app to a PostgreSQL DB
## Instrcutions
From Ioannis
- Determine which is the resource used by TAP to install MySQL or Postgress.
- Add the kubernetes-service-binding extension to your project.
- Add the following config to your project:
```
quarkus.kubernetes-service-binding.services.postgresql.api-version=<api group of the postgress resource>
quarkus.kubernetes-service-binding.services.postgresql.kind=<kind of the postgres resource>
quarkus.kubernetes-service-binding.services.postgresql.name=<name of the postgres resource>
```
- Remove the datasource configuration you have in your application.properies (Just keep the db-kind and optionally name of the datasource)
- Run mvn clean install -Dquarkus.kubernetes.deploy=true
The exact steps I am using are here: https://github.com/iocanel/blog/tree/master/using-quarkus-with-the-service-binding-operator
| 1.0 | Ad dthe steps to bind the quarkus app to a service - ## TODO
Add the steps to bind the Quarkus app to a PostgreSQL DB
## Instrcutions
From Ioannis
- Determine which is the resource used by TAP to install MySQL or Postgress.
- Add the kubernetes-service-binding extension to your project.
- Add the following config to your project:
```
quarkus.kubernetes-service-binding.services.postgresql.api-version=<api group of the postgress resource>
quarkus.kubernetes-service-binding.services.postgresql.kind=<kind of the postgres resource>
quarkus.kubernetes-service-binding.services.postgresql.name=<name of the postgres resource>
```
- Remove the datasource configuration you have in your application.properies (Just keep the db-kind and optionally name of the datasource)
- Run mvn clean install -Dquarkus.kubernetes.deploy=true
The exact steps I am using are here: https://github.com/iocanel/blog/tree/master/using-quarkus-with-the-service-binding-operator
| non_test | ad dthe steps to bind the quarkus app to a service todo add the steps to bind the quarkus app to a postgresql db instrcutions from ioannis determine which is the resource used by tap to install mysql or postgress add the kubernetes service binding extension to your project add the following config to your project quarkus kubernetes service binding services postgresql api version quarkus kubernetes service binding services postgresql kind quarkus kubernetes service binding services postgresql name remove the datasource configuration you have in your application properies just keep the db kind and optionally name of the datasource run mvn clean install dquarkus kubernetes deploy true the exact steps i am using are here | 0 |
315,659 | 27,092,960,100 | IssuesEvent | 2023-02-14 22:54:20 | timescale/timescaledb | https://api.github.com/repos/timescale/timescaledb | opened | [Flaky test] bgw_launcher doesn't see wait_event_changed | flaky-test | ### Which test is flaky?
bgw_launcher
### Since when has the test been flaky?
February 14, 2023
### Link to the failed test run
https://github.com/timescale/timescaledb/actions/runs/4175600633/jobs/7233369995
### Log output
```bash
diff -u /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out
--- /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out 2023-02-14 15:55:10.826147051 +0000
+++ /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out 2023-02-14 15:56:26.894945058 +0000
@@ -142,7 +142,7 @@
AND datname = :'TEST_DBNAME_2';
wait_event_changed
--------------------
- t
+ f
(1 row)
-- Test stop
```
### Reason for flakiness
_No response_ | 1.0 | [Flaky test] bgw_launcher doesn't see wait_event_changed - ### Which test is flaky?
bgw_launcher
### Since when has the test been flaky?
February 14, 2023
### Link to the failed test run
https://github.com/timescale/timescaledb/actions/runs/4175600633/jobs/7233369995
### Log output
```bash
diff -u /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out
--- /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out 2023-02-14 15:55:10.826147051 +0000
+++ /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out 2023-02-14 15:56:26.894945058 +0000
@@ -142,7 +142,7 @@
AND datname = :'TEST_DBNAME_2';
wait_event_changed
--------------------
- t
+ f
(1 row)
-- Test stop
```
### Reason for flakiness
_No response_ | test | bgw launcher doesn t see wait event changed which test is flaky bgw launcher since when has the test been flaky february link to the failed test run log output bash diff u home runner work timescaledb timescaledb test expected bgw launcher out home runner work timescaledb timescaledb build test results bgw launcher out home runner work timescaledb timescaledb test expected bgw launcher out home runner work timescaledb timescaledb build test results bgw launcher out and datname test dbname wait event changed t f row test stop reason for flakiness no response | 1 |
53,466 | 6,330,437,025 | IssuesEvent | 2017-07-26 07:24:25 | jiscdev/study-goal | https://api.github.com/repos/jiscdev/study-goal | closed | Photo incredibly grainy when taken from Android device | Android bug Move to testing Request Accepted | Please can we reduce the compression on the photos - too grainy at the moment.
Photos selected from the library also appear rotated by 90'.
Tested on Samsung S6 physical device.
Photo capture:
<img src="https://user-images.githubusercontent.com/5815497/28213104-8a9777d2-689c-11e7-843a-c119a528ff48.png" width="300">
Library selection:
<img src="https://user-images.githubusercontent.com/5815497/28213110-8c6b48cc-689c-11e7-8d97-9bd3687941b8.png" width="300"> | 1.0 | Photo incredibly grainy when taken from Android device - Please can we reduce the compression on the photos - too grainy at the moment.
Photos selected from the library also appear rotated by 90'.
Tested on Samsung S6 physical device.
Photo capture:
<img src="https://user-images.githubusercontent.com/5815497/28213104-8a9777d2-689c-11e7-843a-c119a528ff48.png" width="300">
Library selection:
<img src="https://user-images.githubusercontent.com/5815497/28213110-8c6b48cc-689c-11e7-8d97-9bd3687941b8.png" width="300"> | test | photo incredibly grainy when taken from android device please can we reduce the compression on the photos too grainy at the moment photos selected from the library also appear rotated by tested on samsung physical device photo capture library selection | 1 |
15,813 | 20,713,349,761 | IssuesEvent | 2022-03-12 08:23:37 | Creators-of-Create/Create | https://api.github.com/repos/Creators-of-Create/Create | closed | JEI potion recipes | bug needs input compatibility stale | clicking on Create potions in jei shows every recipe for automated brewing, item draining, and filling by sprout | True | JEI potion recipes - clicking on Create potions in jei shows every recipe for automated brewing, item draining, and filling by sprout | non_test | jei potion recipes clicking on create potions in jei shows every recipe for automated brewing item draining and filling by sprout | 0 |
62,579 | 17,082,305,010 | IssuesEvent | 2021-07-08 07:24:03 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Regression in ResultQuery.fetchMap(Field, Field) with null values | T: Defect | ### Expected behavior
ResultQuery.fetchMap(Field, Field) should not throw NPE when value field maps to a null value. This was the behavior on 3.14.x. fetchMap's docs don't list any limitations on value field, so I assume this is unintentional. Probably also applies to other fetchMap overrides.
### Actual behavior
NPE is thrown internally from HashMap.merge, at least with Java 16.
java.lang.NullPointerException
at java.base/java.util.HashMap.merge(HashMap.java:1359)
at java.base/java.util.stream.Collectors.lambda$toMap$68(Collectors.java:1666)
at java.base/java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913)
at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682)
at org.jooq.impl.AbstractCursor.collect(AbstractCursor.java:78)
at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:358)
at org.jooq.impl.ResultQueryTrait.fetchMap(ResultQueryTrait.java:850)
### Steps to reproduce the problem
An example query, where TEAM.BUSINESS_ID is a nullable column.
```java
var businessIds = dsl.select()
.from(TEAM)
.fetchMap(TEAM.ID, TEAM.BUSINESS_ID);
```
### Versions
- jOOQ: 3.15.0
- Java: 16
| 1.0 | Regression in ResultQuery.fetchMap(Field, Field) with null values - ### Expected behavior
ResultQuery.fetchMap(Field, Field) should not throw NPE when value field maps to a null value. This was the behavior on 3.14.x. fetchMap's docs don't list any limitations on value field, so I assume this is unintentional. Probably also applies to other fetchMap overrides.
### Actual behavior
NPE is thrown internally from HashMap.merge, at least with Java 16.
java.lang.NullPointerException
at java.base/java.util.HashMap.merge(HashMap.java:1359)
at java.base/java.util.stream.Collectors.lambda$toMap$68(Collectors.java:1666)
at java.base/java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913)
at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682)
at org.jooq.impl.AbstractCursor.collect(AbstractCursor.java:78)
at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:358)
at org.jooq.impl.ResultQueryTrait.fetchMap(ResultQueryTrait.java:850)
### Steps to reproduce the problem
An example query, where TEAM.BUSINESS_ID is a nullable column.
```java
var businessIds = dsl.select()
.from(TEAM)
.fetchMap(TEAM.ID, TEAM.BUSINESS_ID);
```
### Versions
- jOOQ: 3.15.0
- Java: 16
| non_test | regression in resultquery fetchmap field field with null values expected behavior resultquery fetchmap field field should not throw npe when value field maps to a null value this was the behavior on x fetchmap s docs don t list any limitations on value field so i assume this is unintentional probably also applies to other fetchmap overrides actual behavior npe is thrown internally from hashmap merge at least with java java lang nullpointerexception at java base java util hashmap merge hashmap java at java base java util stream collectors lambda tomap collectors java at java base java util stream reduceops accept reduceops java at java base java util iterator foreachremaining iterator java at java base java util spliterators iteratorspliterator foreachremaining spliterators java at java base java util stream abstractpipeline copyinto abstractpipeline java at java base java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java base java util stream reduceops reduceop evaluatesequential reduceops java at java base java util stream abstractpipeline evaluate abstractpipeline java at java base java util stream referencepipeline collect referencepipeline java at org jooq impl abstractcursor collect abstractcursor java at org jooq impl resultquerytrait collect resultquerytrait java at org jooq impl resultquerytrait fetchmap resultquerytrait java steps to reproduce the problem an example query where team business id is a nullable column java var businessids dsl select from team fetchmap team id team business id versions jooq java | 0 |
118,030 | 9,968,483,883 | IssuesEvent | 2019-07-08 15:41:03 | golang/go | https://api.github.com/repos/golang/go | opened | crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder | Soon Testing release-blocker | `crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder.
First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself.
Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6
```
--- FAIL: TestVerifyHostnameResumed (0.10s)
--- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
--- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
FAIL
FAIL crypto/tls 0.828s
```
CC @agl | 1.0 | crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder - `crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder.
First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself.
Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6
```
--- FAIL: TestVerifyHostnameResumed (0.10s)
--- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
--- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
FAIL
FAIL crypto/tls 0.828s
```
CC @agl | test | crypto tls testverifyhostnameresumed consistently failing in longtest builder crypto tls testverifyhostnameresumed is consistently failing in the linux longtest builder first failure was at cl which appears to be unrelated that suggests some sort of change in a non hermetic dependency or perhaps in the builder itself example fail testverifyhostnameresumed fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail fail crypto tls cc agl | 1 |
278,131 | 21,058,099,640 | IssuesEvent | 2022-04-01 06:44:07 | isabelteo/ped | https://api.github.com/repos/isabelteo/ped | opened | Standardising t/Breed | type.DocumentationBug severity.Low |


Can consider to match the command summary, which uses t/Tag with t/Breed which is used above
<!--session: 1648792986020-e42dd825-f9a0-404a-9fc9-346c11c52df9-->
<!--Version: Web v3.4.2--> | 1.0 | Standardising t/Breed -


Can consider to match the command summary, which uses t/Tag with t/Breed which is used above
<!--session: 1648792986020-e42dd825-f9a0-404a-9fc9-346c11c52df9-->
<!--Version: Web v3.4.2--> | non_test | standardising t breed can consider to match the command summary which uses t tag with t breed which is used above | 0 |
79,580 | 7,720,158,160 | IssuesEvent | 2018-05-23 21:54:17 | NMGRL/pychron | https://api.github.com/repos/NMGRL/pychron | closed | table group | Data Reduction Enhancement Ready to close Tested OK | add a table group identifier to the tag file
add set table group tag in same places as set_tag.
use same set_tag code except no need to interact with database. only need to write/commit to tag file
0,0,0,1,1,1
a,b,c,d,e,f
...
s=IntermediateAnalysis(d,e,f)
a,b,c,s
the summary age is the sum of a,b,c,s NOT a,b,c,d,e,f
| 1.0 | table group - add a table group identifier to the tag file
add set table group tag in same places as set_tag.
use same set_tag code except no need to interact with database. only need to write/commit to tag file
0,0,0,1,1,1
a,b,c,d,e,f
...
s=IntermediateAnalysis(d,e,f)
a,b,c,s
the summary age is the sum of a,b,c,s NOT a,b,c,d,e,f
| test | table group add a table group identifier to the tag file add set table group tag in same places as set tag use same set tag code except no need to interact with database only need to write commit to tag file a b c d e f s intermediateanalysis d e f a b c s the summary age is the sum of a b c s not a b c d e f | 1 |
84,053 | 24,209,674,760 | IssuesEvent | 2022-09-25 18:15:37 | contour-terminal/contour | https://api.github.com/repos/contour-terminal/contour | closed | Cannot build with Qt 6.2.4 | build | ### Contour Terminal version
master
### Installer source
Github: source code cloned
### Operating System
Devuan
### Architecture
x86-64
### Other Software
_No response_
### Steps to reproduce
Configure with `-DCONTOUR_BUILD_WITH_QT6=ON` option.
### Expected Behavior
Successfully compiling.
### Actual Behavior
```
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp: In function ‘void BlurBehind::setEnabled(QWindow*, bool, QRegion)’:
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:37:28: error: invalid use of incomplete type ‘class QVariant’
37 | window->setProperty("kwin_blur", region);
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~
In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/qtypeinfo.h:42,
from /usr/include/x86_64-linux-gnu/qt6/QtCore/qglobal.h:1431,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qtguiglobal.h:43,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:43,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/QWindow:1,
from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.h:16,
from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:14:
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:38:28: error: invalid use of incomplete type ‘class QVariant’
38 | window->setProperty("kwin_background_region", region);
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:39:57: error: cannot convert ‘int’ to ‘const QVariant&’
39 | window->setProperty("kwin_background_contrast", 1);
| ^
| |
| int
In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/QObject:1,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:44:
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:40:58: error: cannot convert ‘int’ to ‘const QVariant&’
40 | window->setProperty("kwin_background_intensity", 1);
| ^
| |
| int
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:41:59: error: cannot convert ‘int’ to ‘const QVariant&’
41 | window->setProperty("kwin_background_saturation", 1);
| ^
| |
| int
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:42:46: error: cannot convert ‘const char [38]’ to ‘const QVariant&’
42 | window->setProperty("_MUTTER_HINTS", "blur-provider=sigma:60,brightness:0.9");
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
| |
| const char [38]
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:46:28: error: invalid use of incomplete type ‘const class QVariant’
46 | window->setProperty("kwin_blur", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:47:28: error: invalid use of incomplete type ‘const class QVariant’
47 | window->setProperty("kwin_background_region", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:48:28: error: invalid use of incomplete type ‘const class QVariant’
48 | window->setProperty("kwin_background_contrast", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:49:28: error: invalid use of incomplete type ‘const class QVariant’
49 | window->setProperty("kwin_background_intensity", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:50:28: error: invalid use of incomplete type ‘const class QVariant’
50 | window->setProperty("kwin_background_saturation", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:51:28: error: invalid use of incomplete type ‘const class QVariant’
51 | window->setProperty("_MUTTER_HINTS", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
make[2]: *** [src/contour/CMakeFiles/contour.dir/build.make:140: src/contour/CMakeFiles/contour.dir/BlurBehind.cpp.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [CMakeFiles/Makefile2:556: src/contour/CMakeFiles/contour.dir/all] Error 2
make: *** [Makefile:156: all] Error 2
```
### Additional notes
_No response_ | 1.0 | Cannot build with Qt 6.2.4 - ### Contour Terminal version
master
### Installer source
Github: source code cloned
### Operating System
Devuan
### Architecture
x86-64
### Other Software
_No response_
### Steps to reproduce
Configure with `-DCONTOUR_BUILD_WITH_QT6=ON` option.
### Expected Behavior
Successfully compiling.
### Actual Behavior
```
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp: In function ‘void BlurBehind::setEnabled(QWindow*, bool, QRegion)’:
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:37:28: error: invalid use of incomplete type ‘class QVariant’
37 | window->setProperty("kwin_blur", region);
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~
In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/qtypeinfo.h:42,
from /usr/include/x86_64-linux-gnu/qt6/QtCore/qglobal.h:1431,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qtguiglobal.h:43,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:43,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/QWindow:1,
from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.h:16,
from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:14:
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:38:28: error: invalid use of incomplete type ‘class QVariant’
38 | window->setProperty("kwin_background_region", region);
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:39:57: error: cannot convert ‘int’ to ‘const QVariant&’
39 | window->setProperty("kwin_background_contrast", 1);
| ^
| |
| int
In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/QObject:1,
from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:44:
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:40:58: error: cannot convert ‘int’ to ‘const QVariant&’
40 | window->setProperty("kwin_background_intensity", 1);
| ^
| |
| int
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:41:59: error: cannot convert ‘int’ to ‘const QVariant&’
41 | window->setProperty("kwin_background_saturation", 1);
| ^
| |
| int
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:42:46: error: cannot convert ‘const char [38]’ to ‘const QVariant&’
42 | window->setProperty("_MUTTER_HINTS", "blur-provider=sigma:60,brightness:0.9");
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
| |
| const char [38]
/usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’
376 | bool setProperty(const char *name, const QVariant &value);
| ~~~~~~~~~~~~~~~~^~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:46:28: error: invalid use of incomplete type ‘const class QVariant’
46 | window->setProperty("kwin_blur", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:47:28: error: invalid use of incomplete type ‘const class QVariant’
47 | window->setProperty("kwin_background_region", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:48:28: error: invalid use of incomplete type ‘const class QVariant’
48 | window->setProperty("kwin_background_contrast", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:49:28: error: invalid use of incomplete type ‘const class QVariant’
49 | window->setProperty("kwin_background_intensity", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:50:28: error: invalid use of incomplete type ‘const class QVariant’
50 | window->setProperty("kwin_background_saturation", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
/home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:51:28: error: invalid use of incomplete type ‘const class QVariant’
51 | window->setProperty("_MUTTER_HINTS", {});
| ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~
/usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’
72 | class QVariant;
| ^~~~~~~~
make[2]: *** [src/contour/CMakeFiles/contour.dir/build.make:140: src/contour/CMakeFiles/contour.dir/BlurBehind.cpp.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [CMakeFiles/Makefile2:556: src/contour/CMakeFiles/contour.dir/all] Error 2
make: *** [Makefile:156: all] Error 2
```
### Additional notes
_No response_ | non_test | cannot build with qt contour terminal version master installer source github source code cloned operating system devuan architecture other software no response steps to reproduce configure with dcontour build with on option expected behavior successfully compiling actual behavior home dataman projects c terminals contour src contour blurbehind cpp in function ‘void blurbehind setenabled qwindow bool qregion ’ home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘class qvariant’ window setproperty kwin blur region in file included from usr include linux gnu qtcore qtypeinfo h from usr include linux gnu qtcore qglobal h from usr include linux gnu qtgui qtguiglobal h from usr include linux gnu qtgui qwindow h from usr include linux gnu qtgui qwindow from home dataman projects c terminals contour src contour blurbehind h from home dataman projects c terminals contour src contour blurbehind cpp usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘class qvariant’ window setproperty kwin background region region usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background contrast int in file included from usr include linux gnu qtcore qobject from usr include linux gnu qtgui qwindow h usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background intensity int usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background saturation int usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘const char ’ to ‘const qvariant ’ window setproperty mutter hints blur provider sigma brightness const char usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin blur usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background region usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background contrast usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background intensity usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background saturation usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty mutter hints usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant make error make waiting for unfinished jobs make error make error additional notes no response | 0 |
194,033 | 14,667,706,651 | IssuesEvent | 2020-12-29 19:19:28 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | rkand4/vuejs: pkg/http/listener_test.go; 40 LoC | fresh small test |
Found a possible issue in [rkand4/vuejs](https://www.github.com/rkand4/vuejs) at [pkg/http/listener_test.go](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable i used in defer or goroutine at line 522
[Click here to see the code in its original context.](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543)
<details>
<summary>Click here to show the 40 line(s) of Go which triggered the analyzer.</summary>
```go
for i, testCase := range testCases {
listener, err := newHTTPListener(
testCase.serverAddrs,
testCase.tlsConfig,
time.Duration(0),
time.Duration(0),
time.Duration(0),
nil,
nil,
errorFunc,
)
if err != nil {
t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err)
}
go func() {
serverConn, aerr := listener.Accept()
if aerr == nil {
t.Fatalf("Test %d: accept: expected = <error>, got = <nil>", i+1)
}
if serverConn != nil {
t.Fatalf("Test %d: accept: server expected = <nil>, got = %v", i+1, serverConn)
}
}()
for _, serverAddr := range testCase.serverAddrs {
conn, err := net.Dial("tcp", serverAddr)
if err != nil {
t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err)
}
if _, err = io.WriteString(conn, testCase.request); err != nil {
t.Fatalf("Test %d: request send: expected = <nil>, got = %v", i+1, err)
}
conn.Close()
}
listener.Close()
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: e790bf122760e516384839f20608d7b79b1d344a
| 1.0 | rkand4/vuejs: pkg/http/listener_test.go; 40 LoC -
Found a possible issue in [rkand4/vuejs](https://www.github.com/rkand4/vuejs) at [pkg/http/listener_test.go](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable i used in defer or goroutine at line 522
[Click here to see the code in its original context.](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543)
<details>
<summary>Click here to show the 40 line(s) of Go which triggered the analyzer.</summary>
```go
for i, testCase := range testCases {
listener, err := newHTTPListener(
testCase.serverAddrs,
testCase.tlsConfig,
time.Duration(0),
time.Duration(0),
time.Duration(0),
nil,
nil,
errorFunc,
)
if err != nil {
t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err)
}
go func() {
serverConn, aerr := listener.Accept()
if aerr == nil {
t.Fatalf("Test %d: accept: expected = <error>, got = <nil>", i+1)
}
if serverConn != nil {
t.Fatalf("Test %d: accept: server expected = <nil>, got = %v", i+1, serverConn)
}
}()
for _, serverAddr := range testCase.serverAddrs {
conn, err := net.Dial("tcp", serverAddr)
if err != nil {
t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err)
}
if _, err = io.WriteString(conn, testCase.request); err != nil {
t.Fatalf("Test %d: request send: expected = <nil>, got = %v", i+1, err)
}
conn.Close()
}
listener.Close()
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: e790bf122760e516384839f20608d7b79b1d344a
| test | vuejs pkg http listener test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable i used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for i testcase range testcases listener err newhttplistener testcase serveraddrs testcase tlsconfig time duration time duration time duration nil nil errorfunc if err nil t fatalf test d error expected got v i err go func serverconn aerr listener accept if aerr nil t fatalf test d accept expected got i if serverconn nil t fatalf test d accept server expected got v i serverconn for serveraddr range testcase serveraddrs conn err net dial tcp serveraddr if err nil t fatalf test d error expected got v i err if err io writestring conn testcase request err nil t fatalf test d request send expected got v i err conn close listener close leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
44,518 | 18,069,191,714 | IssuesEvent | 2021-09-20 23:25:07 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Search] Use new elasticsearch-js client in data plugin search service | Feature:Search enhancement loe:hours Team:AppServices impact:low | Related: https://github.com/elastic/kibana/issues/35508
Now that https://github.com/elastic/kibana/pull/69905 is merged, we have access to the new elasticsearch-js client (which updates Typescript support for ES as well as a handful of other benefits), we should update the data plugin search service to use the new client rather than the legacy client. This would open up the ability for other enhancements (see https://github.com/elastic/kibana/issues/70675). | 1.0 | [Search] Use new elasticsearch-js client in data plugin search service - Related: https://github.com/elastic/kibana/issues/35508
Now that https://github.com/elastic/kibana/pull/69905 is merged, we have access to the new elasticsearch-js client (which updates Typescript support for ES as well as a handful of other benefits), we should update the data plugin search service to use the new client rather than the legacy client. This would open up the ability for other enhancements (see https://github.com/elastic/kibana/issues/70675). | non_test | use new elasticsearch js client in data plugin search service related now that is merged we have access to the new elasticsearch js client which updates typescript support for es as well as a handful of other benefits we should update the data plugin search service to use the new client rather than the legacy client this would open up the ability for other enhancements see | 0 |
221,907 | 17,376,543,935 | IssuesEvent | 2021-07-30 22:36:17 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | No error info prompts when switching proxy setting from 'Do not use proxy' to 'Use app proxy settings' without configuring the correct proxy | :beetle: regression :gear: proxy :heavy_check_mark: merged 🧪 testing | **Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210723.3
**Branch**: main
**Platform/OS**: Windows 10
**Architecture**: ia32
**How Found**: From running test cases
**Regression From**: Previous release (1.20.1)
## Steps to Reproduce ##
1. Launch Storage Explorer -> Make sure the proxy setting is 'Do not use proxy' -> Close Storage Explorer.
2. Download Fiddler.
3. Configure Fiddler from https://docs.microsoft.com/en-us/azure/storage/common/storage-explorer-troubleshooting?tabs=Windows%2C2004#network-logs.
4. Open Storage Explorer -> Switch to 'Use app proxy settings'.
5. Check whether an error info bar prompts.
## Expected Experience ##
An error info bar prompts.

## Actual Experience ##
No error info bar prompts.
## Additional Context ##
There is no traffic for operations like 'upload, delete blobs...' after configuring the correct settings for 'Use app proxy settings'. | 1.0 | No error info prompts when switching proxy setting from 'Do not use proxy' to 'Use app proxy settings' without configuring the correct proxy - **Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210723.3
**Branch**: main
**Platform/OS**: Windows 10
**Architecture**: ia32
**How Found**: From running test cases
**Regression From**: Previous release (1.20.1)
## Steps to Reproduce ##
1. Launch Storage Explorer -> Make sure the proxy setting is 'Do not use proxy' -> Close Storage Explorer.
2. Download Fiddler.
3. Configure Fiddler from https://docs.microsoft.com/en-us/azure/storage/common/storage-explorer-troubleshooting?tabs=Windows%2C2004#network-logs.
4. Open Storage Explorer -> Switch to 'Use app proxy settings'.
5. Check whether an error info bar prompts.
## Expected Experience ##
An error info bar prompts.

## Actual Experience ##
No error info bar prompts.
## Additional Context ##
There is no traffic for operations like 'upload, delete blobs...' after configuring the correct settings for 'Use app proxy settings'. | test | no error info prompts when switching proxy setting from do not use proxy to use app proxy settings without configuring the correct proxy storage explorer version dev build number branch main platform os windows architecture how found from running test cases regression from previous release steps to reproduce launch storage explorer make sure the proxy setting is do not use proxy close storage explorer download fiddler configure fiddler from open storage explorer switch to use app proxy settings check whether an error info bar prompts expected experience an error info bar prompts actual experience no error info bar prompts additional context there is no traffic for operations like upload delete blobs after configuring the correct settings for use app proxy settings | 1 |
103,531 | 16,602,910,551 | IssuesEvent | 2021-06-01 22:15:11 | gms-ws-sandbox/nibrs | https://api.github.com/repos/gms-ws-sandbox/nibrs | opened | CVE-2018-1272 (High) detected in spring-core-4.3.11.RELEASE.jar | security vulnerability | ## CVE-2018-1272 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-4.3.11.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.11.RELEASE/spring-core-4.3.11.RELEASE.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/spring-core-4.3.11.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-core-4.3.11.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.
<p>Publish Date: 2018-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272>CVE-2018-1272</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2018-1272">https://tanzu.vmware.com/security/cve-2018-1272</a></p>
<p>Release Date: 2018-04-06</p>
<p>Fix Resolution: org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"4.3.11.RELEASE","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:4.3.11.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-1272 (High) detected in spring-core-4.3.11.RELEASE.jar - ## CVE-2018-1272 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-4.3.11.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.11.RELEASE/spring-core-4.3.11.RELEASE.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/spring-core-4.3.11.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-core-4.3.11.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.
<p>Publish Date: 2018-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272>CVE-2018-1272</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2018-1272">https://tanzu.vmware.com/security/cve-2018-1272</a></p>
<p>Release Date: 2018-04-06</p>
<p>Fix Resolution: org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"4.3.11.RELEASE","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:4.3.11.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in spring core release jar cve high severity vulnerability vulnerable library spring core release jar spring core library home page a href path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library home wss scanner repository org springframework spring core release spring core release jar nibrs tools nibrs fbi service target nibrs fbi service web inf lib spring core release jar dependency hierarchy x spring core release jar vulnerable library found in head commit a href found in base branch master vulnerability details spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release release org springframework spring web release release isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org springframework spring core release isminimumfixversionavailable true minimumfixversion org springframework spring core release release org springframework spring web release release basebranches vulnerabilityidentifier cve vulnerabilitydetails spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles vulnerabilityurl | 0 |
188,921 | 14,479,842,662 | IssuesEvent | 2020-12-10 10:21:02 | Tencent/bk-ci | https://api.github.com/repos/Tencent/bk-ci | closed | 触发构建时,支持填入构建信息 | area/ci/api area/ci/backend area/ci/frontend kind/feat/product priority/critical-urgent stage/test stage/uat test/passed | ### 期望
手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。
如果没有填入,平台给构建信息置入默认值:“手动触发”
后续将作为能替代构建号的重要信息作为构建索引。

### 注意
需考虑以下场景:
代码库触发 - 用最新commit msg置入构建信息;
远程触发 - `curl`指令内需包含`BK_CI_BUILD_MSG`。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值;
定时触发 - 默认“定时触发”;
子流水线触发 - 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入${BK_CI_BUILD_MSG},则替换默认值;
服务触发 - 默认“服务触发”
关联需求:#2850 | 2.0 | 触发构建时,支持填入构建信息 - ### 期望
手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。
如果没有填入,平台给构建信息置入默认值:“手动触发”
后续将作为能替代构建号的重要信息作为构建索引。

### 注意
需考虑以下场景:
代码库触发 - 用最新commit msg置入构建信息;
远程触发 - `curl`指令内需包含`BK_CI_BUILD_MSG`。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值;
定时触发 - 默认“定时触发”;
子流水线触发 - 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入${BK_CI_BUILD_MSG},则替换默认值;
服务触发 - 默认“服务触发”
关联需求:#2850 | test | 触发构建时,支持填入构建信息 期望 手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。 如果没有填入,平台给构建信息置入默认值:“手动触发” 后续将作为能替代构建号的重要信息作为构建索引。 注意 需考虑以下场景: 代码库触发 用最新commit msg置入构建信息; 远程触发 curl 指令内需包含 bk ci build msg 。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值; 定时触发 默认“定时触发”; 子流水线触发 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入 bk ci build msg ,则替换默认值; 服务触发 默认“服务触发” 关联需求: | 1 |
117,437 | 9,935,111,014 | IssuesEvent | 2019-07-02 15:48:05 | somaliaims/Somali-AIMSUI | https://api.github.com/repos/somaliaims/Somali-AIMSUI | closed | sector mapping on input add sector | july testing | The goal of the AIMS is to store all projects with a Somali sector - all the other sector types are just there to assist users entering data that comes from another sector type.
So when the sector is imported from IATI, or another AIMS project, we automatically select the sector type that it is coming from.
When it is a manual entry - we allow the user to select the sector type they want to enter.
In both cases, the UI then needs to show what default (in this case Somali) sector the input sector type will be stored as (and allow the user to adjust if they want to).
If there is no mapping, then the user should be asked to select a Somali sector.
The user then confirms the Somali sector that gets stored. | 1.0 | sector mapping on input add sector - The goal of the AIMS is to store all projects with a Somali sector - all the other sector types are just there to assist users entering data that comes from another sector type.
So when the sector is imported from IATI, or another AIMS project, we automatically select the sector type that it is coming from.
When it is a manual entry - we allow the user to select the sector type they want to enter.
In both cases, the UI then needs to show what default (in this case Somali) sector the input sector type will be stored as (and allow the user to adjust if they want to).
If there is no mapping, then the user should be asked to select a Somali sector.
The user then confirms the Somali sector that gets stored. | test | sector mapping on input add sector the goal of the aims is to store all projects with a somali sector all the other sector types are just there to assist users entering data that comes from another sector type so when the sector is imported from iati or another aims project we automatically select the sector type that it is coming from when it is a manual entry we allow the user to select the sector type they want to enter in both cases the ui then needs to show what default in this case somali sector the input sector type will be stored as and allow the user to adjust if they want to if there is no mapping then the user should be asked to select a somali sector the user then confirms the somali sector that gets stored | 1 |
104,954 | 9,013,179,094 | IssuesEvent | 2019-02-05 18:48:30 | NillerMedDild/Enigmatica2Expert | https://api.github.com/repos/NillerMedDild/Enigmatica2Expert | closed | Fluids not Despawning properly after source block removed | needs testing | I've noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal.....in particular the pool water from the slime islands and the healing spa water from the alps biomes.
## Thank you for helping us improve. #####
_Please follow the issue template, and do not delete any part of it._
_Failure to comply will result in the issue being closed._
**Issue Description**
**Modpack version (Do not use "latest")**
**Logs**
If applicable, please upload your crash report, or latest.log to [gist](https://gist.github.com/).
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
| 1.0 | Fluids not Despawning properly after source block removed - I've noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal.....in particular the pool water from the slime islands and the healing spa water from the alps biomes.
## Thank you for helping us improve. #####
_Please follow the issue template, and do not delete any part of it._
_Failure to comply will result in the issue being closed._
**Issue Description**
**Modpack version (Do not use "latest")**
**Logs**
If applicable, please upload your crash report, or latest.log to [gist](https://gist.github.com/).
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
| test | fluids not despawning properly after source block removed i ve noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal in particular the pool water from the slime islands and the healing spa water from the alps biomes thank you for helping us improve please follow the issue template and do not delete any part of it failure to comply will result in the issue being closed issue description modpack version do not use latest logs if applicable please upload your crash report or latest log to screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here | 1 |
285,399 | 24,663,808,692 | IssuesEvent | 2022-10-18 08:47:36 | rancher/cis-operator | https://api.github.com/repos/rancher/cis-operator | closed | ClusterRoles for cis-operator and scan job pods are too permissive | [zube]: To Test team/area3 team/infracloud feature/charts-cis-benchmark team/rke2 | ## Issue
The existing ClusteRoles ([here](https://github.com/rancher/charts/blob/dev-v2.6/charts/rancher-cis-benchmark/2.1.0/templates/rbac.yaml)) attached to the ServiceAccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported. For instance:
```
groups:
- id: exploitable-controle
text: ""
checks:
- id: exploitable-check
text: ""
audit: >
kubectl delete ns kube-system
```
## Potential Solution
In the Helm chart, the roles attached to the SAs should be scope down to specific required functions. We've been testing and these roles seem to work pretty well for all existing scans + custom scans:
#### Operator ClusterRole and Role
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: cis-operator-clusterrole
rules:
- apiGroups:
- "cis.cattle.io"
resources:
- "*"
verbs:
- "*"
- apiGroups:
- ""
resources:
- "pods"
- "services"
- "configmaps"
- "nodes"
verbs:
- "list"
- "create"
- "update"
- "watch"
- apiGroups:
- "batch"
resources:
- "jobs"
verbs:
- "list"
- "create"
- "update"
- "watch"
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: cis-operator-role
namespace: cis-operator-system
rules:
- apiGroups:
- ""
resources:
- "services"
verbs:
- "watch"
- "list"
- "get"
- "patch"
- apiGroups:
- "batch"
resources:
- "jobs"
verbs:
- "watch"
- "list"
- "get"
- "delete"
- apiGroups:
- ""
resources:
- "configmaps"
- "pods"
verbs:
- "watch"
- "list"
- "get"
- "delete"
```
#### Scan ClusterRole & Role
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: cis-scan-clusterrole
rules:
- apiGroups:
- "*"
resources:
- "*"
verbs:
- "list"
- "describe"
- "get"
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: cis-scan-role
namespace: cis-operator-system
rules:
- apiGroups:
- ""
resources:
- "pods"
verbs:
- "patch"
- "get"
- "list"
- "describe"
- apiGroups:
- "apps"
resources:
- "daemonsets"
verbs:
- "*"
- apiGroups:
- ""
resources:
- "secrets"
- "configmaps"
verbs:
- "update"
- "create"
```
| 1.0 | ClusterRoles for cis-operator and scan job pods are too permissive - ## Issue
The existing ClusteRoles ([here](https://github.com/rancher/charts/blob/dev-v2.6/charts/rancher-cis-benchmark/2.1.0/templates/rbac.yaml)) attached to the ServiceAccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported. For instance:
```
groups:
- id: exploitable-controle
text: ""
checks:
- id: exploitable-check
text: ""
audit: >
kubectl delete ns kube-system
```
## Potential Solution
In the Helm chart, the roles attached to the SAs should be scope down to specific required functions. We've been testing and these roles seem to work pretty well for all existing scans + custom scans:
#### Operator ClusterRole and Role
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: cis-operator-clusterrole
rules:
- apiGroups:
- "cis.cattle.io"
resources:
- "*"
verbs:
- "*"
- apiGroups:
- ""
resources:
- "pods"
- "services"
- "configmaps"
- "nodes"
verbs:
- "list"
- "create"
- "update"
- "watch"
- apiGroups:
- "batch"
resources:
- "jobs"
verbs:
- "list"
- "create"
- "update"
- "watch"
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: cis-operator-role
namespace: cis-operator-system
rules:
- apiGroups:
- ""
resources:
- "services"
verbs:
- "watch"
- "list"
- "get"
- "patch"
- apiGroups:
- "batch"
resources:
- "jobs"
verbs:
- "watch"
- "list"
- "get"
- "delete"
- apiGroups:
- ""
resources:
- "configmaps"
- "pods"
verbs:
- "watch"
- "list"
- "get"
- "delete"
```
#### Scan ClusterRole & Role
```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: cis-scan-clusterrole
rules:
- apiGroups:
- "*"
resources:
- "*"
verbs:
- "list"
- "describe"
- "get"
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: cis-scan-role
namespace: cis-operator-system
rules:
- apiGroups:
- ""
resources:
- "pods"
verbs:
- "patch"
- "get"
- "list"
- "describe"
- apiGroups:
- "apps"
resources:
- "daemonsets"
verbs:
- "*"
- apiGroups:
- ""
resources:
- "secrets"
- "configmaps"
verbs:
- "update"
- "create"
```
| test | clusterroles for cis operator and scan job pods are too permissive issue the existing clusteroles attached to the serviceaccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported for instance groups id exploitable controle text checks id exploitable check text audit kubectl delete ns kube system potential solution in the helm chart the roles attached to the sas should be scope down to specific required functions we ve been testing and these roles seem to work pretty well for all existing scans custom scans operator clusterrole and role yaml apiversion rbac authorization io kind clusterrole metadata name cis operator clusterrole rules apigroups cis cattle io resources verbs apigroups resources pods services configmaps nodes verbs list create update watch apigroups batch resources jobs verbs list create update watch apiversion rbac authorization io kind role metadata name cis operator role namespace cis operator system rules apigroups resources services verbs watch list get patch apigroups batch resources jobs verbs watch list get delete apigroups resources configmaps pods verbs watch list get delete scan clusterrole role yaml apiversion rbac authorization io kind clusterrole metadata name cis scan clusterrole rules apigroups resources verbs list describe get apiversion rbac authorization io kind role metadata name cis scan role namespace cis operator system rules apigroups resources pods verbs patch get list describe apigroups apps resources daemonsets verbs apigroups resources secrets configmaps verbs update create | 1 |
140,179 | 5,398,323,310 | IssuesEvent | 2017-02-27 16:40:51 | metacpan/metacpan-web | https://api.github.com/repos/metacpan/metacpan-web | closed | Allow users to view their user_id | priority:Low Volunteer needed | Please let the site's users view their user id, so that they will then be able to ask 3rd-party sites & services to fetch their MetaCPAN favorites or interact in other ways.
One such service is https://www.perlmodules.net/, where retrieving one's own list of favorites would be quite useful.
So, exposing the user id under https://metacpan.org/account/profile might be the most appropriate place.
| 1.0 | Allow users to view their user_id - Please let the site's users view their user id, so that they will then be able to ask 3rd-party sites & services to fetch their MetaCPAN favorites or interact in other ways.
One such service is https://www.perlmodules.net/, where retrieving one's own list of favorites would be quite useful.
So, exposing the user id under https://metacpan.org/account/profile might be the most appropriate place.
| non_test | allow users to view their user id please let the site s users view their user id so that they will then be able to ask party sites services to fetch their metacpan favorites or interact in other ways one such service is where retrieving one s own list of favorites would be quite useful so exposing the user id under might be the most appropriate place | 0 |
57,888 | 6,559,861,391 | IssuesEvent | 2017-09-07 06:53:43 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | OSX: intermittent failure: async-hooks/test-callback-error | async_hooks macos test | * **Version**: master
* **Platform**: OSX
* **Subsystem**: async hooks
<!-- Enter your issue details below this comment. -->
Failure in run for unrelated change:
https://ci.nodejs.org/job/node-test-commit-osx/12183/nodes=osx1010/console
```
not ok 76 async-hooks/test-callback-error
---
duration_ms: 15.551
severity: fail
stack: |-
start case 1
end case 1: 114.761ms
start case 2
end case 2: 115.167ms
start case 3
end case 3: 8.221ms
Error: test_callback_abort
at ActivityCollector.initHooks.oninit.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:36:45)
at ActivityCollector.oninit (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:509:15)
at ActivityCollector._init (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/init-hooks.js:182:10)
at emitInitNative (async_hooks.js:446:43)
at Object.emitInitScript [as emitInit] (async_hooks.js:349:3)
at Object.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:38:17)
at Module._compile (module.js:549:30)
at Object.Module._extensions..js (module.js:560:10)
at Module.load (module.js:483:32)
at tryModuleLoad (module.js:446:12)
1: node::Abort() [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
6: 0x31be80046fd
``` | 1.0 | OSX: intermittent failure: async-hooks/test-callback-error - * **Version**: master
* **Platform**: OSX
* **Subsystem**: async hooks
<!-- Enter your issue details below this comment. -->
Failure in run for unrelated change:
https://ci.nodejs.org/job/node-test-commit-osx/12183/nodes=osx1010/console
```
not ok 76 async-hooks/test-callback-error
---
duration_ms: 15.551
severity: fail
stack: |-
start case 1
end case 1: 114.761ms
start case 2
end case 2: 115.167ms
start case 3
end case 3: 8.221ms
Error: test_callback_abort
at ActivityCollector.initHooks.oninit.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:36:45)
at ActivityCollector.oninit (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:509:15)
at ActivityCollector._init (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/init-hooks.js:182:10)
at emitInitNative (async_hooks.js:446:43)
at Object.emitInitScript [as emitInit] (async_hooks.js:349:3)
at Object.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:38:17)
at Module._compile (module.js:549:30)
at Object.Module._extensions..js (module.js:560:10)
at Module.load (module.js:483:32)
at tryModuleLoad (module.js:446:12)
1: node::Abort() [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node]
6: 0x31be80046fd
``` | test | osx intermittent failure async hooks test callback error version master platform osx subsystem async hooks failure in run for unrelated change not ok async hooks test callback error duration ms severity fail stack start case end case start case end case start case end case error test callback abort at activitycollector inithooks oninit common mustcall users iojs build workspace node test commit osx nodes test async hooks test callback error js at activitycollector oninit users iojs build workspace node test commit osx nodes test common index js at activitycollector init users iojs build workspace node test commit osx nodes test async hooks init hooks js at emitinitnative async hooks js at object emitinitscript async hooks js at object users iojs build workspace node test commit osx nodes test async hooks test callback error js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js node abort node chdir functioncallbackinfo const internal functioncallbackarguments call void functioncallbackinfo const internal maybehandle internal anonymous namespace handleapicallhelper internal isolate internal handle internal handle internal handle internal handle internal builtinarguments internal builtin impl handleapicall internal builtinarguments internal isolate | 1 |
309,052 | 26,649,126,881 | IssuesEvent | 2023-01-25 12:22:17 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql/tests: TestRandomSyntaxGeneration failed | C-test-failure O-robot branch-master T-sql-sessions | sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=artifacts#/) on master @ [642afd6e8c8d1f967da4aa0c3c08e2bdc0495100](https://github.com/cockroachdb/cockroach/commits/642afd6e8c8d1f967da4aa0c3c08e2bdc0495100):
```
crypto/tls.(*atLeastReader).Read(0xc05661a8b8, {0xc04a583d40?, 0x0?, 0x389488168619053b?})
GOROOT/src/crypto/tls/conn.go:787 +0x3d fp=0xc0339d0938 sp=0xc0339d08f0 pc=0x86799d
bytes.(*Buffer).ReadFrom(0xc0569a3078, {0x6c2e660, 0xc05661a8b8})
GOROOT/src/bytes/buffer.go:202 +0x98 fp=0xc0339d0990 sp=0xc0339d0938 pc=0x565858
crypto/tls.(*Conn).readFromUntil(0xc0569a2e00, {0x6c3dd20?, 0xc04bd20b18}, 0x80deb1fe72be5d74?)
GOROOT/src/crypto/tls/conn.go:809 +0xe5 fp=0xc0339d09d0 sp=0xc0339d0990 pc=0x867b85
crypto/tls.(*Conn).readRecordOrCCS(0xc0569a2e00, 0x0)
GOROOT/src/crypto/tls/conn.go:616 +0x116 fp=0xc0339d0d58 sp=0xc0339d09d0 pc=0x864fd6
crypto/tls.(*Conn).readRecord(...)
GOROOT/src/crypto/tls/conn.go:582
crypto/tls.(*Conn).readHandshake(0xc0569a2e00)
GOROOT/src/crypto/tls/conn.go:1019 +0x6d fp=0xc0339d0e38 sp=0xc0339d0d58 pc=0x868ded
crypto/tls.(*serverHandshakeStateTLS13).readClientCertificate(0xc0339d11c0)
GOROOT/src/crypto/tls/handshake_server_tls13.go:787 +0x65 fp=0xc0339d10c0 sp=0xc0339d0e38 pc=0x8948e5
crypto/tls.(*serverHandshakeStateTLS13).handshake(0xc0339d11c0)
GOROOT/src/crypto/tls/handshake_server_tls13.go:78 +0xb6 fp=0xc0339d10e0 sp=0xc0339d10c0 pc=0x890bb6
crypto/tls.(*Conn).serverHandshake(0xc0569a2e00, {0x6c75730, 0xc0569a0b00})
GOROOT/src/crypto/tls/handshake_server.go:54 +0xd0 fp=0xc0339d12b8 sp=0xc0339d10e0 pc=0x88b550
crypto/tls.(*Conn).serverHandshake-fm({0x6c75730?, 0xc0569a0b00?})
<autogenerated>:1 +0x39 fp=0xc0339d12e0 sp=0xc0339d12b8 pc=0x89f4d9
crypto/tls.(*Conn).handshakeContext(0xc0569a2e00, {0x6c75768, 0xc000080048})
GOROOT/src/crypto/tls/conn.go:1462 +0x32f fp=0xc0339d14a8 sp=0xc0339d12e0 pc=0x86b6cf
crypto/tls.(*Conn).HandshakeContext(...)
GOROOT/src/crypto/tls/conn.go:1405
crypto/tls.(*Conn).Handshake(...)
GOROOT/src/crypto/tls/conn.go:1389
crypto/tls.(*Conn).Read(0xc0569a2e00, {0xc0569c4438, 0x4, 0x493f10?})
GOROOT/src/crypto/tls/conn.go:1274 +0x65 fp=0xc0339d1518 sp=0xc0339d14a8 pc=0x86ab25
io.ReadAtLeast({0x7eff005e27f8, 0xc0569a2e00}, {0xc0569c4438, 0x4, 0x4}, 0x4)
GOROOT/src/io/io.go:332 +0x9a fp=0xc0339d1560 sp=0xc0339d1518 pc=0x50ccba
io.ReadFull(...)
GOROOT/src/io/io.go:351
github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase.(*ReadBuffer).ReadUntypedMsg(0xc0569c4420, {0x7eff005e27f8, 0xc0569a2e00})
github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase/encoding.go:155 +0x65 fp=0xc0339d1720 sp=0xc0339d1560 pc=0x27e6c85
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).readVersion(0xc000cf23c0, {0x7eff005e27f8, 0xc0569a2e00})
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1371 +0x127 fp=0xc0339d17a8 sp=0xc0339d1720 pc=0x38ce567
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).maybeUpgradeToSecureConn(0xc000cf23c0, {0x0?, 0xffc?}, {0x6cbb9a8, 0xc04bd20b18}, 0x2, 0x6c3dd20?, 0xc0339d1bc0)
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1315 +0x35d fp=0xc0339d1878 sp=0xc0339d17a8 pc=0x38cdd9d
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).ServeConn(0xc000cf23c0, {0x6c75730, 0xc056b06880}, {0x6cbb9a8?, 0xc04bd20b18}, 0x6a?)
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:811 +0x59e fp=0xc0339d1e20 sp=0xc0339d1878 pc=0x38ca83e
github.com/cockroachdb/cockroach/pkg/server.(*SQLServer).startServeSQL.func1.1({0x6c757d8?, 0xc0568207b0?}, {0x6cbb9a8, 0xc04bd20b18})
github.com/cockroachdb/cockroach/pkg/server/server_sql.go:1568 +0xc8 fp=0xc0339d1eb0 sp=0xc0339d1e20 pc=0x3b95648
github.com/cockroachdb/cockroach/pkg/util/netutil.(*TCPServer).ServeWith.func1({0x6c757d8, 0xc0568207b0})
github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:188 +0xc7 fp=0xc0339d1f30 sp=0xc0339d1eb0 pc=0x17d96a7
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2()
github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 +0x146 fp=0xc0339d1fe0 sp=0xc0339d1f30 pc=0x10fdae6
runtime.goexit()
GOROOT/src/runtime/asm_amd64.s:1594 +0x1 fp=0xc0339d1fe8 sp=0xc0339d1fe0 pc=0x4c4021
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx
github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:461 +0x43b
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #89363 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.2.0]
- #87572 sql/tests: TestRandomSyntaxGeneration failed [DROP OWNED BY timeout] [C-test-failure O-robot T-sql-schema branch-release-22.2]
- #77893 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.1]
- #74271 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.2]
- #65210 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.1]
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23617 | 1.0 | sql/tests: TestRandomSyntaxGeneration failed - sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=artifacts#/) on master @ [642afd6e8c8d1f967da4aa0c3c08e2bdc0495100](https://github.com/cockroachdb/cockroach/commits/642afd6e8c8d1f967da4aa0c3c08e2bdc0495100):
```
crypto/tls.(*atLeastReader).Read(0xc05661a8b8, {0xc04a583d40?, 0x0?, 0x389488168619053b?})
GOROOT/src/crypto/tls/conn.go:787 +0x3d fp=0xc0339d0938 sp=0xc0339d08f0 pc=0x86799d
bytes.(*Buffer).ReadFrom(0xc0569a3078, {0x6c2e660, 0xc05661a8b8})
GOROOT/src/bytes/buffer.go:202 +0x98 fp=0xc0339d0990 sp=0xc0339d0938 pc=0x565858
crypto/tls.(*Conn).readFromUntil(0xc0569a2e00, {0x6c3dd20?, 0xc04bd20b18}, 0x80deb1fe72be5d74?)
GOROOT/src/crypto/tls/conn.go:809 +0xe5 fp=0xc0339d09d0 sp=0xc0339d0990 pc=0x867b85
crypto/tls.(*Conn).readRecordOrCCS(0xc0569a2e00, 0x0)
GOROOT/src/crypto/tls/conn.go:616 +0x116 fp=0xc0339d0d58 sp=0xc0339d09d0 pc=0x864fd6
crypto/tls.(*Conn).readRecord(...)
GOROOT/src/crypto/tls/conn.go:582
crypto/tls.(*Conn).readHandshake(0xc0569a2e00)
GOROOT/src/crypto/tls/conn.go:1019 +0x6d fp=0xc0339d0e38 sp=0xc0339d0d58 pc=0x868ded
crypto/tls.(*serverHandshakeStateTLS13).readClientCertificate(0xc0339d11c0)
GOROOT/src/crypto/tls/handshake_server_tls13.go:787 +0x65 fp=0xc0339d10c0 sp=0xc0339d0e38 pc=0x8948e5
crypto/tls.(*serverHandshakeStateTLS13).handshake(0xc0339d11c0)
GOROOT/src/crypto/tls/handshake_server_tls13.go:78 +0xb6 fp=0xc0339d10e0 sp=0xc0339d10c0 pc=0x890bb6
crypto/tls.(*Conn).serverHandshake(0xc0569a2e00, {0x6c75730, 0xc0569a0b00})
GOROOT/src/crypto/tls/handshake_server.go:54 +0xd0 fp=0xc0339d12b8 sp=0xc0339d10e0 pc=0x88b550
crypto/tls.(*Conn).serverHandshake-fm({0x6c75730?, 0xc0569a0b00?})
<autogenerated>:1 +0x39 fp=0xc0339d12e0 sp=0xc0339d12b8 pc=0x89f4d9
crypto/tls.(*Conn).handshakeContext(0xc0569a2e00, {0x6c75768, 0xc000080048})
GOROOT/src/crypto/tls/conn.go:1462 +0x32f fp=0xc0339d14a8 sp=0xc0339d12e0 pc=0x86b6cf
crypto/tls.(*Conn).HandshakeContext(...)
GOROOT/src/crypto/tls/conn.go:1405
crypto/tls.(*Conn).Handshake(...)
GOROOT/src/crypto/tls/conn.go:1389
crypto/tls.(*Conn).Read(0xc0569a2e00, {0xc0569c4438, 0x4, 0x493f10?})
GOROOT/src/crypto/tls/conn.go:1274 +0x65 fp=0xc0339d1518 sp=0xc0339d14a8 pc=0x86ab25
io.ReadAtLeast({0x7eff005e27f8, 0xc0569a2e00}, {0xc0569c4438, 0x4, 0x4}, 0x4)
GOROOT/src/io/io.go:332 +0x9a fp=0xc0339d1560 sp=0xc0339d1518 pc=0x50ccba
io.ReadFull(...)
GOROOT/src/io/io.go:351
github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase.(*ReadBuffer).ReadUntypedMsg(0xc0569c4420, {0x7eff005e27f8, 0xc0569a2e00})
github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase/encoding.go:155 +0x65 fp=0xc0339d1720 sp=0xc0339d1560 pc=0x27e6c85
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).readVersion(0xc000cf23c0, {0x7eff005e27f8, 0xc0569a2e00})
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1371 +0x127 fp=0xc0339d17a8 sp=0xc0339d1720 pc=0x38ce567
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).maybeUpgradeToSecureConn(0xc000cf23c0, {0x0?, 0xffc?}, {0x6cbb9a8, 0xc04bd20b18}, 0x2, 0x6c3dd20?, 0xc0339d1bc0)
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1315 +0x35d fp=0xc0339d1878 sp=0xc0339d17a8 pc=0x38cdd9d
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).ServeConn(0xc000cf23c0, {0x6c75730, 0xc056b06880}, {0x6cbb9a8?, 0xc04bd20b18}, 0x6a?)
github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:811 +0x59e fp=0xc0339d1e20 sp=0xc0339d1878 pc=0x38ca83e
github.com/cockroachdb/cockroach/pkg/server.(*SQLServer).startServeSQL.func1.1({0x6c757d8?, 0xc0568207b0?}, {0x6cbb9a8, 0xc04bd20b18})
github.com/cockroachdb/cockroach/pkg/server/server_sql.go:1568 +0xc8 fp=0xc0339d1eb0 sp=0xc0339d1e20 pc=0x3b95648
github.com/cockroachdb/cockroach/pkg/util/netutil.(*TCPServer).ServeWith.func1({0x6c757d8, 0xc0568207b0})
github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:188 +0xc7 fp=0xc0339d1f30 sp=0xc0339d1eb0 pc=0x17d96a7
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2()
github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 +0x146 fp=0xc0339d1fe0 sp=0xc0339d1f30 pc=0x10fdae6
runtime.goexit()
GOROOT/src/runtime/asm_amd64.s:1594 +0x1 fp=0xc0339d1fe8 sp=0xc0339d1fe0 pc=0x4c4021
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx
github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:461 +0x43b
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #89363 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.2.0]
- #87572 sql/tests: TestRandomSyntaxGeneration failed [DROP OWNED BY timeout] [C-test-failure O-robot T-sql-schema branch-release-22.2]
- #77893 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.1]
- #74271 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.2]
- #65210 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.1]
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23617 | test | sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration with on master crypto tls atleastreader read goroot src crypto tls conn go fp sp pc bytes buffer readfrom goroot src bytes buffer go fp sp pc crypto tls conn readfromuntil goroot src crypto tls conn go fp sp pc crypto tls conn readrecordorccs goroot src crypto tls conn go fp sp pc crypto tls conn readrecord goroot src crypto tls conn go crypto tls conn readhandshake goroot src crypto tls conn go fp sp pc crypto tls readclientcertificate goroot src crypto tls handshake server go fp sp pc crypto tls handshake goroot src crypto tls handshake server go fp sp pc crypto tls conn serverhandshake goroot src crypto tls handshake server go fp sp pc crypto tls conn serverhandshake fm fp sp pc crypto tls conn handshakecontext goroot src crypto tls conn go fp sp pc crypto tls conn handshakecontext goroot src crypto tls conn go crypto tls conn handshake goroot src crypto tls conn go crypto tls conn read goroot src crypto tls conn go fp sp pc io readatleast goroot src io io go fp sp pc io readfull goroot src io io go github com cockroachdb cockroach pkg sql pgwire pgwirebase readbuffer readuntypedmsg github com cockroachdb cockroach pkg sql pgwire pgwirebase encoding go fp sp pc github com cockroachdb cockroach pkg sql pgwire server readversion github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg sql pgwire server maybeupgradetosecureconn github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg sql pgwire server serveconn github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg server sqlserver startservesql github com cockroachdb cockroach pkg server server sql go fp sp pc github com cockroachdb cockroach pkg util netutil tcpserver servewith github com cockroachdb cockroach pkg util netutil net go fp sp pc github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go fp sp pc runtime goexit goroot src runtime asm s fp sp pc created by github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go help see also same failure on other branches sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed cc cockroachdb sql sessions jira issue crdb | 1 |
43,974 | 5,579,233,275 | IssuesEvent | 2017-03-28 14:13:14 | Microsoft/vscode | https://api.github.com/repos/Microsoft/vscode | closed | Test: new commands to increase/decrease view size | testplan-item | Refs: https://github.com/Microsoft/vscode/issues/22645
- [x] anyOS - @weinand (macOS)
- [x] anyOS - @egamma
New commands (`workbench.action.increaseViewSize` and `workbench.action.decreaseViewSize`) allow to increase or decrease a view (sidebar, editor or panel) size with keyboard only.
Verify
* the commands work as advertised on the view that has current keyboard focus
* you cannot break the layout (e.g. minimal and maximal dimensions are respected) | 1.0 | Test: new commands to increase/decrease view size - Refs: https://github.com/Microsoft/vscode/issues/22645
- [x] anyOS - @weinand (macOS)
- [x] anyOS - @egamma
New commands (`workbench.action.increaseViewSize` and `workbench.action.decreaseViewSize`) allow to increase or decrease a view (sidebar, editor or panel) size with keyboard only.
Verify
* the commands work as advertised on the view that has current keyboard focus
* you cannot break the layout (e.g. minimal and maximal dimensions are respected) | test | test new commands to increase decrease view size refs anyos weinand macos anyos egamma new commands workbench action increaseviewsize and workbench action decreaseviewsize allow to increase or decrease a view sidebar editor or panel size with keyboard only verify the commands work as advertised on the view that has current keyboard focus you cannot break the layout e g minimal and maximal dimensions are respected | 1 |
516,307 | 14,979,063,594 | IssuesEvent | 2021-01-28 11:42:40 | Uninett/Argus-frontend | https://api.github.com/repos/Uninett/Argus-frontend | opened | Bug: login background isn't reset to paper after login when using userpass-auth | bug polish priority: low size: small usability | **Describe the bug**
The login view sets the background using the useBackground() hook, but never reset's it. When the user logs using FEIDE this isn't a problem because FEIDE authentication causes a page reload. When the user logs in using userpass-auth however, it will NOT cause a reload, and the background stays cyan.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to /login
2. Enter username and password
3. Press enter
4. Observe that the background used in the login view is still set
**Expected behavior**
The background should be reset to the default for material ui, which is paper
**Screenshots**

**Solution**
Use the useEffect hook's cleanup property to unset the background. This could
probably be done in the useBackground hook.
| 1.0 | Bug: login background isn't reset to paper after login when using userpass-auth - **Describe the bug**
The login view sets the background using the useBackground() hook, but never reset's it. When the user logs using FEIDE this isn't a problem because FEIDE authentication causes a page reload. When the user logs in using userpass-auth however, it will NOT cause a reload, and the background stays cyan.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to /login
2. Enter username and password
3. Press enter
4. Observe that the background used in the login view is still set
**Expected behavior**
The background should be reset to the default for material ui, which is paper
**Screenshots**

**Solution**
Use the useEffect hook's cleanup property to unset the background. This could
probably be done in the useBackground hook.
| non_test | bug login background isn t reset to paper after login when using userpass auth describe the bug the login view sets the background using the usebackground hook but never reset s it when the user logs using feide this isn t a problem because feide authentication causes a page reload when the user logs in using userpass auth however it will not cause a reload and the background stays cyan to reproduce steps to reproduce the behavior go to login enter username and password press enter observe that the background used in the login view is still set expected behavior the background should be reset to the default for material ui which is paper screenshots solution use the useeffect hook s cleanup property to unset the background this could probably be done in the usebackground hook | 0 |
331,908 | 29,170,456,303 | IssuesEvent | 2023-05-19 00:58:12 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | DISABLED test_build_tuple_unpack_dynamic_shapes_static_default (__main__.StaticDefaultDynamicShapesMiscTests) | triaged module: flaky-tests skipped module: dynamo | Platforms: asan, linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack_dynamic_shapes_static_default&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_build_tuple_unpack_dynamic_shapes_static_default`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py` | 1.0 | DISABLED test_build_tuple_unpack_dynamic_shapes_static_default (__main__.StaticDefaultDynamicShapesMiscTests) - Platforms: asan, linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack_dynamic_shapes_static_default&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_build_tuple_unpack_dynamic_shapes_static_default`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py` | test | disabled test build tuple unpack dynamic shapes static default main staticdefaultdynamicshapesmisctests platforms asan linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test build tuple unpack dynamic shapes static default there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path dynamo test dynamic shapes py or dynamo test dynamic shapes py | 1 |
116,960 | 9,904,496,122 | IssuesEvent | 2019-06-27 09:18:29 | HarmanU/test-utility-app | https://api.github.com/repos/HarmanU/test-utility-app | opened | Test Plan PDF generator | Test Plan Creator enhancement | Exports a PDF for the test plan
can use similar method to the Jira card exporter to added customisation | 1.0 | Test Plan PDF generator - Exports a PDF for the test plan
can use similar method to the Jira card exporter to added customisation | test | test plan pdf generator exports a pdf for the test plan can use similar method to the jira card exporter to added customisation | 1 |
256,081 | 22,036,520,069 | IssuesEvent | 2022-05-28 17:17:14 | Cookie-AutoDelete/Cookie-AutoDelete | https://api.github.com/repos/Cookie-AutoDelete/Cookie-AutoDelete | opened | [Bug] LocalStorage is cleaned even when "Keep LocalStorage" is enabled for a site | untested bug/issue | ### Acknowledgements
- [X] I acknowledge that I have read the above items
### Describe the bug
[Assuming I've understood things correctly](https://github.com/Cookie-AutoDelete/Cookie-AutoDelete/issues/982#issuecomment-1140297632), there issue I'm having is this: With LocalStorage enabled in the Settings under the _Other Browsing Data Cleanup_ options and also _Keep LocalStorage_ ticked for a particular site's whitelist options, every time I start the browser, the LocalStorage is cleaned. At least, this is all I can assume because even though LocalStorage is _not_ mentioned in the logs after restarting the browser, and even with everything ticked for the site's options, the only way to keep myself signed in is to _disable_ LocalStorage in the global settings.
The site is https://mail.proton.me and the rule in use is `*.proton.me` (which should cover account.proton.me and other related/necessary sites to maintain being logged-in). All the expression's "Keep" options are enabled.
### To Reproduce
1. Enable LocalStorage Cleanup in the global settings.
2. Add `*.proton.me` expression, tick every "Keep" option.
3. Log-in to the service, wait till it loads, close the browser.
4. Re-open the browser and be asked to sign in again.
### Expected Behavior
Step 4 should not ask you to sign in again, i.e. keep the LocalStorage when _Keep LocalStorage_ is ticked for the expression.
### Screenshots
_No response_
### System Info - Operating System (OS)
Windows 11
### System Info - Browser Info
Brave 1.39.111 Chromium: 102.0.5005.61 (Official Build) (64-bit)
### System Info - CookieAutoDelete Version
3.7.0
### Additional Context
_No response_ | 1.0 | [Bug] LocalStorage is cleaned even when "Keep LocalStorage" is enabled for a site - ### Acknowledgements
- [X] I acknowledge that I have read the above items
### Describe the bug
[Assuming I've understood things correctly](https://github.com/Cookie-AutoDelete/Cookie-AutoDelete/issues/982#issuecomment-1140297632), there issue I'm having is this: With LocalStorage enabled in the Settings under the _Other Browsing Data Cleanup_ options and also _Keep LocalStorage_ ticked for a particular site's whitelist options, every time I start the browser, the LocalStorage is cleaned. At least, this is all I can assume because even though LocalStorage is _not_ mentioned in the logs after restarting the browser, and even with everything ticked for the site's options, the only way to keep myself signed in is to _disable_ LocalStorage in the global settings.
The site is https://mail.proton.me and the rule in use is `*.proton.me` (which should cover account.proton.me and other related/necessary sites to maintain being logged-in). All the expression's "Keep" options are enabled.
### To Reproduce
1. Enable LocalStorage Cleanup in the global settings.
2. Add `*.proton.me` expression, tick every "Keep" option.
3. Log-in to the service, wait till it loads, close the browser.
4. Re-open the browser and be asked to sign in again.
### Expected Behavior
Step 4 should not ask you to sign in again, i.e. keep the LocalStorage when _Keep LocalStorage_ is ticked for the expression.
### Screenshots
_No response_
### System Info - Operating System (OS)
Windows 11
### System Info - Browser Info
Brave 1.39.111 Chromium: 102.0.5005.61 (Official Build) (64-bit)
### System Info - CookieAutoDelete Version
3.7.0
### Additional Context
_No response_ | test | localstorage is cleaned even when keep localstorage is enabled for a site acknowledgements i acknowledge that i have read the above items describe the bug there issue i m having is this with localstorage enabled in the settings under the other browsing data cleanup options and also keep localstorage ticked for a particular site s whitelist options every time i start the browser the localstorage is cleaned at least this is all i can assume because even though localstorage is not mentioned in the logs after restarting the browser and even with everything ticked for the site s options the only way to keep myself signed in is to disable localstorage in the global settings the site is and the rule in use is proton me which should cover account proton me and other related necessary sites to maintain being logged in all the expression s keep options are enabled to reproduce enable localstorage cleanup in the global settings add proton me expression tick every keep option log in to the service wait till it loads close the browser re open the browser and be asked to sign in again expected behavior step should not ask you to sign in again i e keep the localstorage when keep localstorage is ticked for the expression screenshots no response system info operating system os windows system info browser info brave chromium official build bit system info cookieautodelete version additional context no response | 1 |
29,776 | 4,535,896,269 | IssuesEvent | 2016-09-08 18:44:15 | appium/appium | https://api.github.com/repos/appium/appium | opened | /wd/hub/session/:sessionId/appium/device/lock not implemented | iOS XCUITest | `/wd/hub/session/:sessionId/appium/device/lock` not implemented in xcuitest driver | 1.0 | /wd/hub/session/:sessionId/appium/device/lock not implemented - `/wd/hub/session/:sessionId/appium/device/lock` not implemented in xcuitest driver | test | wd hub session sessionid appium device lock not implemented wd hub session sessionid appium device lock not implemented in xcuitest driver | 1 |
766,064 | 26,870,308,981 | IssuesEvent | 2023-02-04 11:30:11 | AUBGTheHUB/spa-website-2022 | https://api.github.com/repos/AUBGTheHUB/spa-website-2022 | closed | Footer | high priority api frontend SPA HackAUBG | ## Brief description:
Design and create the Footer component
## User Story
Should have instagram, facebook, email, address? .
Should be reusable in HackAUBG and other pages -> a prop should be passed which would either render OG design (used in main page and Jobs page) or Hack design (difference is in the background color).
you can probably set a property with a default value (e.g. `usedInHackPage = false`). Contact me @asynchroza after developing the initial footer and we will decide together how to approach this issue.
| 1.0 | Footer - ## Brief description:
Design and create the Footer component
## User Story
Should have instagram, facebook, email, address? .
Should be reusable in HackAUBG and other pages -> a prop should be passed which would either render OG design (used in main page and Jobs page) or Hack design (difference is in the background color).
you can probably set a property with a default value (e.g. `usedInHackPage = false`). Contact me @asynchroza after developing the initial footer and we will decide together how to approach this issue.
| non_test | footer brief description design and create the footer component user story should have instagram facebook email address should be reusable in hackaubg and other pages a prop should be passed which would either render og design used in main page and jobs page or hack design difference is in the background color you can probably set a property with a default value e g usedinhackpage false contact me asynchroza after developing the initial footer and we will decide together how to approach this issue | 0 |
303,392 | 26,204,302,721 | IssuesEvent | 2023-01-03 20:46:27 | BoBAdministration/QA-Bug-Reports | https://api.github.com/repos/BoBAdministration/QA-Bug-Reports | closed | Emote wheel default chats display two "Yes" options | can't reproduce Fixed-PendingTesting | **Describe the Bug**
In the wheel, you should have a yes and no option, instead, both options both say "yes".
**To Reproduce**
Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones.
1. Log in to any server
2. Spawn as anything
3. Use the emote wheel to get to the chat buttons
4. Both options will say yes, or both options will say yes & no. However, both options will say yes.
**Expected behavior**
There would be a yes and no option.
**Actual behavior**
Both say yes.
**Screenshots & Video**

https://user-images.githubusercontent.com/48891088/210025796-e48eaf52-775f-4021-be2f-725fde43910c.mp4
**Branch Version**
Live & Tester
**Character Information**
Paras 1.2
**Additional Information**
Titania | 1.0 | Emote wheel default chats display two "Yes" options - **Describe the Bug**
In the wheel, you should have a yes and no option, instead, both options both say "yes".
**To Reproduce**
Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones.
1. Log in to any server
2. Spawn as anything
3. Use the emote wheel to get to the chat buttons
4. Both options will say yes, or both options will say yes & no. However, both options will say yes.
**Expected behavior**
There would be a yes and no option.
**Actual behavior**
Both say yes.
**Screenshots & Video**

https://user-images.githubusercontent.com/48891088/210025796-e48eaf52-775f-4021-be2f-725fde43910c.mp4
**Branch Version**
Live & Tester
**Character Information**
Paras 1.2
**Additional Information**
Titania | test | emote wheel default chats display two yes options describe the bug in the wheel you should have a yes and no option instead both options both say yes to reproduce steps to reproduce the behavior in detail please include all steps even menial ones log in to any server spawn as anything use the emote wheel to get to the chat buttons both options will say yes or both options will say yes no however both options will say yes expected behavior there would be a yes and no option actual behavior both say yes screenshots video branch version live tester character information paras additional information titania | 1 |
108,024 | 9,256,946,129 | IssuesEvent | 2019-03-17 00:06:06 | rstoneback/pysat | https://api.github.com/repos/rstoneback/pysat | opened | Minimize code duplication in testing instruments | testing | There are several areas in the testing instruments that can be combined to reduce duplication of code.
- `testaddx` instruments can be folded into the main pysat_testing.py instrument, using tags to change data output.
- `testsmall` instruments can be potentially be folded back into pysat_testing.py, as `tag='100'` would create default behaviour.
- With current parameters, the test constellations `test_diff`, `test_diff2`, and `test_diff_small` should perform the exact same function. @rstoneback, is this a placeholder for future tests, or can these be combined? | 1.0 | Minimize code duplication in testing instruments - There are several areas in the testing instruments that can be combined to reduce duplication of code.
- `testaddx` instruments can be folded into the main pysat_testing.py instrument, using tags to change data output.
- `testsmall` instruments can be potentially be folded back into pysat_testing.py, as `tag='100'` would create default behaviour.
- With current parameters, the test constellations `test_diff`, `test_diff2`, and `test_diff_small` should perform the exact same function. @rstoneback, is this a placeholder for future tests, or can these be combined? | test | minimize code duplication in testing instruments there are several areas in the testing instruments that can be combined to reduce duplication of code testaddx instruments can be folded into the main pysat testing py instrument using tags to change data output testsmall instruments can be potentially be folded back into pysat testing py as tag would create default behaviour with current parameters the test constellations test diff test and test diff small should perform the exact same function rstoneback is this a placeholder for future tests or can these be combined | 1 |
79,187 | 15,586,110,645 | IssuesEvent | 2021-03-18 01:11:49 | mibo32/fitbit-api-example-java | https://api.github.com/repos/mibo32/fitbit-api-example-java | opened | CVE-2020-24750 (High) detected in jackson-databind-2.8.1.jar | security vulnerability | ## CVE-2020-24750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: fitbit-api-example-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-28</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-24750 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2020-24750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: fitbit-api-example-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-28</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file fitbit api example java pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
148,547 | 11,856,388,624 | IssuesEvent | 2020-03-25 07:22:55 | onaio/reveal-frontend | https://api.github.com/repos/onaio/reveal-frontend | opened | Test blocks with jest spyon lead to test failure if they appear above other tests. | tests | In `/src/containers/pages/FocusInvestigation/map/active/tests/index.test.tsx`
move the last test block (the one with @todo comment) to the top and run the tests again. They will fail | 1.0 | Test blocks with jest spyon lead to test failure if they appear above other tests. - In `/src/containers/pages/FocusInvestigation/map/active/tests/index.test.tsx`
move the last test block (the one with @todo comment) to the top and run the tests again. They will fail | test | test blocks with jest spyon lead to test failure if they appear above other tests in src containers pages focusinvestigation map active tests index test tsx move the last test block the one with todo comment to the top and run the tests again they will fail | 1 |
279,461 | 24,227,110,307 | IssuesEvent | 2022-09-26 15:10:28 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] XPackRestIT test {p0=transform/transforms_unattended/Test unattended put and start wildcard} failing | >test-failure :ml/Transform Team:ML | **Build scan:**
https://gradle-enterprise.elastic.co/s/mc6mzbewge5rq/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_unattended%2FTest%20unattended%20put%20and%20start%20wildcard%7D
**Reproduction line:**
`./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_unattended/Test unattended put and start wildcard}" -Dtests.seed=EF18978E741FE47A -Dtests.locale=no -Dtests.timezone=Antarctica/Vostok -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_unattended/Test%20unattended%20put%20and%20start%20wildcard%7D
**Failure excerpt:**
```
java.lang.AssertionError: Failure at [transform/transforms_unattended:14]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484)
at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.junit.Assert.fail(Assert.java:88)
at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484)
at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | 1.0 | [CI] XPackRestIT test {p0=transform/transforms_unattended/Test unattended put and start wildcard} failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/mc6mzbewge5rq/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_unattended%2FTest%20unattended%20put%20and%20start%20wildcard%7D
**Reproduction line:**
`./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_unattended/Test unattended put and start wildcard}" -Dtests.seed=EF18978E741FE47A -Dtests.locale=no -Dtests.timezone=Antarctica/Vostok -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_unattended/Test%20unattended%20put%20and%20start%20wildcard%7D
**Failure excerpt:**
```
java.lang.AssertionError: Failure at [transform/transforms_unattended:14]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484)
at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.junit.Assert.fail(Assert.java:88)
at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484)
at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | test | xpackrestit test transform transforms unattended test unattended put and start wildcard failing build scan reproduction line gradlew x pack plugin yamlresttest tests org elasticsearch xpack test rest xpackrestit test transform transforms unattended test unattended put and start wildcard dtests seed dtests locale no dtests timezone antarctica vostok druntime java applicable branches main reproduces locally no failure history failure excerpt java lang assertionerror failure at expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect invoke null at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java caused by java lang assertionerror expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at org junit assert fail assert java at org elasticsearch test rest yaml section dosection execute dosection java at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect invoke null at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 1 |
301,732 | 26,092,644,629 | IssuesEvent | 2022-12-26 13:53:17 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Failing test: Jest Tests.x-pack/plugins/cases/public/components/all_cases - AllCasesListGeneric Actions Row actions should disable row actions when selecting a case | failed-test | A test failed on a tracked branch
```
TestingLibraryElementError: Unable to find an element by: [data-test-subj="checkboxSelectRow-basic-case-id"]
Ignored nodes: comments, script, style
<body
class=""
>
<div />
</body>
at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19
at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1010:31)
at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12)
at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1009:12)
at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10)
at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40)
at runMicrotasks (<anonymous>)
at runNextTicks (node:internal/process/task_queues:61:5)
at listOnTimeout (node:internal/timers:528:9)
at processTimers (node:internal/timers:502:7)
at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3)
at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25182#01854e8e-ad88-4aeb-95ef-5e047b0791fa)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/all_cases","test.name":"AllCasesListGeneric Actions Row actions should disable row actions when selecting a case","test.failCount":1}} --> | 1.0 | Failing test: Jest Tests.x-pack/plugins/cases/public/components/all_cases - AllCasesListGeneric Actions Row actions should disable row actions when selecting a case - A test failed on a tracked branch
```
TestingLibraryElementError: Unable to find an element by: [data-test-subj="checkboxSelectRow-basic-case-id"]
Ignored nodes: comments, script, style
<body
class=""
>
<div />
</body>
at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19
at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1010:31)
at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12)
at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1009:12)
at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10)
at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40)
at runMicrotasks (<anonymous>)
at runNextTicks (node:internal/process/task_queues:61:5)
at listOnTimeout (node:internal/timers:528:9)
at processTimers (node:internal/timers:502:7)
at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3)
at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25182#01854e8e-ad88-4aeb-95ef-5e047b0791fa)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/all_cases","test.name":"AllCasesListGeneric Actions Row actions should disable row actions when selecting a case","test.failCount":1}} --> | test | failing test jest tests x pack plugins cases public components all cases allcaseslistgeneric actions row actions should disable row actions when selecting a case a test failed on a tracked branch testinglibraryelementerror unable to find an element by ignored nodes comments script style body class at object getelementerror var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist config js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at getbytestid var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components all cases all cases list test tsx at batchedupdates var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom development js at act var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom test utils development js at object var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components all cases all cases list test tsx at promise then completed var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at new promise at callasynccircusfn var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at callcircustest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runmicrotasks at runnextticks node internal process task queues at listontimeout node internal timers at processtimers node internal timers at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at run var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runandtransformresultstojestformat var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js first failure | 1 |
451,490 | 13,036,859,590 | IssuesEvent | 2020-07-28 12:57:51 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1636] Web elections: graphs don't match | Category: Elections Website Category: Web Priority: High Status: Fixed | 1. This is how graph looks in the preview for comments

See, this one has bars for actions
2. After saving we don't have this bar here no more

| 1.0 | [0.9.0 staging-1636] Web elections: graphs don't match - 1. This is how graph looks in the preview for comments

See, this one has bars for actions
2. After saving we don't have this bar here no more

| non_test | web elections graphs don t match this is how graph looks in the preview for comments see this one has bars for actions after saving we don t have this bar here no more | 0 |
342,873 | 30,642,309,636 | IssuesEvent | 2023-07-24 23:35:09 | microsoft/vscode-remote-release | https://api.github.com/repos/microsoft/vscode-remote-release | opened | Test: local port forwarding using tunnels | testplan-item | Refs https://github.com/microsoft/vscode-internalbacklog/issues/4382
- [ ] macOS
- [ ] windows
Complexity: 3
Authors: @connor4312, @alexr00
---
This iteration we built the first version of "local forwarding", which allows users to forward their local ports and make them accessible to the internet using Dev Tunnels.
1. Install the latest prerelease of the Remote Tunnels extension (note: this is [still in PR](https://github.com/microsoft/vscode-remote-tunnels/pull/661) at time of writing, so may not be published by Zurich's Tuesday morning)
2. Set `"remote.tunnels.experimental.forwarding": true` in your user settings, and reload the window
3. The "Ports" view should now appear, even though you are not connected to a remote.
4. You should be able to forward a port and go to the resulting URL to view anything you have running on that port locally.
5. Test out forwarding multiple ports from multiple VS Code windows.
6. Capture the Output from the "Remote Tunnels" extension if you have any issues | 1.0 | Test: local port forwarding using tunnels - Refs https://github.com/microsoft/vscode-internalbacklog/issues/4382
- [ ] macOS
- [ ] windows
Complexity: 3
Authors: @connor4312, @alexr00
---
This iteration we built the first version of "local forwarding", which allows users to forward their local ports and make them accessible to the internet using Dev Tunnels.
1. Install the latest prerelease of the Remote Tunnels extension (note: this is [still in PR](https://github.com/microsoft/vscode-remote-tunnels/pull/661) at time of writing, so may not be published by Zurich's Tuesday morning)
2. Set `"remote.tunnels.experimental.forwarding": true` in your user settings, and reload the window
3. The "Ports" view should now appear, even though you are not connected to a remote.
4. You should be able to forward a port and go to the resulting URL to view anything you have running on that port locally.
5. Test out forwarding multiple ports from multiple VS Code windows.
6. Capture the Output from the "Remote Tunnels" extension if you have any issues | test | test local port forwarding using tunnels refs macos windows complexity authors this iteration we built the first version of local forwarding which allows users to forward their local ports and make them accessible to the internet using dev tunnels install the latest prerelease of the remote tunnels extension note this is at time of writing so may not be published by zurich s tuesday morning set remote tunnels experimental forwarding true in your user settings and reload the window the ports view should now appear even though you are not connected to a remote you should be able to forward a port and go to the resulting url to view anything you have running on that port locally test out forwarding multiple ports from multiple vs code windows capture the output from the remote tunnels extension if you have any issues | 1 |
4,238 | 20,999,654,346 | IssuesEvent | 2022-03-29 16:13:25 | jxk20/nlb_goodreads_searcher | https://api.github.com/repos/jxk20/nlb_goodreads_searcher | closed | Set up github CI/CD pipeline | maintainability | - [x] Do tests for `client`
- [x] Check for code coverage
- [ ] Do tests for `server` | True | Set up github CI/CD pipeline - - [x] Do tests for `client`
- [x] Check for code coverage
- [ ] Do tests for `server` | non_test | set up github ci cd pipeline do tests for client check for code coverage do tests for server | 0 |
319,204 | 27,355,908,428 | IssuesEvent | 2023-02-27 12:51:52 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | lorax-templates v4.2.1 (r4.2) | r4.2-host-cur-test | Update of lorax-templates to v4.2.1 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-lorax-templates/commit/969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799
[Changes since previous version](https://github.com/QubesOS/qubes-lorax-templates/compare/v4.2.0...v4.2.1):
QubesOS/qubes-lorax-templates@969b6dc version 4.2.1
QubesOS/qubes-lorax-templates@5e0324f Handle kernel versions with .qubes.fc suffix, instead of .fc.qubes
QubesOS/qubes-lorax-templates@df7326a Use inst.repo= instead of inst.stage2=
QubesOS/qubes-lorax-templates@19dd3fe Unbreak optional sshd
QubesOS/qubes-lorax-templates@81c0646 Install artwork package
QubesOS/qubes-lorax-templates@4debf10 Clarify kernel-latest boot entry
Referenced issues:
QubesOS/qubes-issues#6792
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current all` (available 5 days from now)
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
| 1.0 | lorax-templates v4.2.1 (r4.2) - Update of lorax-templates to v4.2.1 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-lorax-templates/commit/969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799
[Changes since previous version](https://github.com/QubesOS/qubes-lorax-templates/compare/v4.2.0...v4.2.1):
QubesOS/qubes-lorax-templates@969b6dc version 4.2.1
QubesOS/qubes-lorax-templates@5e0324f Handle kernel versions with .qubes.fc suffix, instead of .fc.qubes
QubesOS/qubes-lorax-templates@df7326a Use inst.repo= instead of inst.stage2=
QubesOS/qubes-lorax-templates@19dd3fe Unbreak optional sshd
QubesOS/qubes-lorax-templates@81c0646 Install artwork package
QubesOS/qubes-lorax-templates@4debf10 Clarify kernel-latest boot entry
Referenced issues:
QubesOS/qubes-issues#6792
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current all` (available 5 days from now)
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
| test | lorax templates update of lorax templates to for qubes see comments below for details and build status from commit qubesos qubes lorax templates version qubesos qubes lorax templates handle kernel versions with qubes fc suffix instead of fc qubes qubesos qubes lorax templates use inst repo instead of inst qubesos qubes lorax templates unbreak optional sshd qubesos qubes lorax templates install artwork package qubesos qubes lorax templates clarify kernel latest boot entry referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload component lorax templates current all available days from now upload component lorax templates security testing you can choose subset of distributions like upload component lorax templates current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at | 1 |
127,779 | 10,482,223,270 | IssuesEvent | 2019-09-24 11:31:57 | chainer/chainer | https://api.github.com/repos/chainer/chainer | closed | flaky test: `chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction::test_double_backward` | cat:test pr-ongoing prio:high | Possibly related to #7946
https://jenkins.preferred.jp/job/chainer/job/cupy_pr/347/TEST=chainer-py35,label=mn1-p100/console
>`FAIL tests/chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always_param_25_{dilate=1, nobias=False, x_dtype=float16, contiguous=None, W_dtype=float16, groups=2, cover_all=False}::test_double_backward`
```
18:48:25 E chainer.testing.function_link.FunctionTestError: Parameterized test failed.
18:48:25 E
18:48:25 E Base test method: TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always.test_double_backward
18:48:25 E Test parameters:
18:48:25 E dilate: 1
18:48:25 E nobias: False
18:48:25 E x_dtype: <class 'numpy.float16'>
18:48:25 E contiguous: None
18:48:25 E W_dtype: <class 'numpy.float16'>
18:48:25 E groups: 2
18:48:25 E cover_all: False
18:48:25 E
18:48:25 E
18:48:25 E (caused by)
18:48:25 E FunctionTestError: double backward is not implemented correctly
18:48:25 E
18:48:25 E (caused by)
18:48:25 E AssertionError: check_double_backward failed (eps=0.001 atol=0.001 rtol=0.001)
...
18:48:25 E gradients (numeric): -0.020542602051870057
18:48:25 E gradients (backward): -0.01935833355256278
18:48:25 E
18:48:25 E x: numeric gradient, y: backward gradient
18:48:25 E Not equal to tolerance rtol=0.001, atol=0.001
18:48:25 E
18:48:25 E (mismatch 100.0%)
18:48:25 E x: array(-0.020542602051870057)
18:48:25 E y: array(-0.01935833355256278)
18:48:25 E
18:48:25 E assert_allclose failed:
18:48:25 E shape: () ()
18:48:25 E dtype: float64 float64
18:48:25 E i: (0,)
18:48:25 E x[i]: -0.020542602051870057
18:48:25 E y[i]: -0.01935833355256278
18:48:25 E relative error[i]: 0.061176159409160366
18:48:25 E absolute error[i]: 0.0011842684993072783
18:48:25 E relative tolerance * |y[i]|: 1.9358333552562778e-05
18:48:25 E absolute tolerance: 0.001
18:48:25 E total tolerance: 0.001019358333552563
18:48:25 E x: -0.020542602051870057
18:48:25 E y: -0.01935833355256278
``` | 1.0 | flaky test: `chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction::test_double_backward` - Possibly related to #7946
https://jenkins.preferred.jp/job/chainer/job/cupy_pr/347/TEST=chainer-py35,label=mn1-p100/console
>`FAIL tests/chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always_param_25_{dilate=1, nobias=False, x_dtype=float16, contiguous=None, W_dtype=float16, groups=2, cover_all=False}::test_double_backward`
```
18:48:25 E chainer.testing.function_link.FunctionTestError: Parameterized test failed.
18:48:25 E
18:48:25 E Base test method: TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always.test_double_backward
18:48:25 E Test parameters:
18:48:25 E dilate: 1
18:48:25 E nobias: False
18:48:25 E x_dtype: <class 'numpy.float16'>
18:48:25 E contiguous: None
18:48:25 E W_dtype: <class 'numpy.float16'>
18:48:25 E groups: 2
18:48:25 E cover_all: False
18:48:25 E
18:48:25 E
18:48:25 E (caused by)
18:48:25 E FunctionTestError: double backward is not implemented correctly
18:48:25 E
18:48:25 E (caused by)
18:48:25 E AssertionError: check_double_backward failed (eps=0.001 atol=0.001 rtol=0.001)
...
18:48:25 E gradients (numeric): -0.020542602051870057
18:48:25 E gradients (backward): -0.01935833355256278
18:48:25 E
18:48:25 E x: numeric gradient, y: backward gradient
18:48:25 E Not equal to tolerance rtol=0.001, atol=0.001
18:48:25 E
18:48:25 E (mismatch 100.0%)
18:48:25 E x: array(-0.020542602051870057)
18:48:25 E y: array(-0.01935833355256278)
18:48:25 E
18:48:25 E assert_allclose failed:
18:48:25 E shape: () ()
18:48:25 E dtype: float64 float64
18:48:25 E i: (0,)
18:48:25 E x[i]: -0.020542602051870057
18:48:25 E y[i]: -0.01935833355256278
18:48:25 E relative error[i]: 0.061176159409160366
18:48:25 E absolute error[i]: 0.0011842684993072783
18:48:25 E relative tolerance * |y[i]|: 1.9358333552562778e-05
18:48:25 E absolute tolerance: 0.001
18:48:25 E total tolerance: 0.001019358333552563
18:48:25 E x: -0.020542602051870057
18:48:25 E y: -0.01935833355256278
``` | test | flaky test chainer tests functions tests connection tests test convolution py test double backward possibly related to fail tests chainer tests functions tests connection tests test convolution py use chainerx false chainerx device none use cuda false cuda device none use cudnn never cudnn deterministic false autotune false cudnn fast batch normalization false use ideep always param dilate nobias false x dtype contiguous none w dtype groups cover all false test double backward e chainer testing function link functiontesterror parameterized test failed e e base test method use chainerx false chainerx device none use cuda false cuda device none use cudnn never cudnn deterministic false autotune false cudnn fast batch normalization false use ideep always test double backward e test parameters e dilate e nobias false e x dtype e contiguous none e w dtype e groups e cover all false e e e caused by e functiontesterror double backward is not implemented correctly e e caused by e assertionerror check double backward failed eps atol rtol e gradients numeric e gradients backward e e x numeric gradient y backward gradient e not equal to tolerance rtol atol e e mismatch e x array e y array e e assert allclose failed e shape e dtype e i e x e y e relative error e absolute error e relative tolerance y e absolute tolerance e total tolerance e x e y | 1 |
23,679 | 4,031,735,207 | IssuesEvent | 2016-05-18 18:11:03 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | circleci: failed tests: TestTxnPutOutOfOrder | Robot test-failure | The following test appears to have failed:
[#18040](https://circleci.com/gh/cockroachdb/cockroach/18040):
```
I160518 17:35:26.740337 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740471 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740604 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740729 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740859 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
panic: test timed out after 1m10s
goroutine 14692 [running]:
panic(0x156f960, 0xc821c80050)
/usr/local/go/src/runtime/panic.go:464 +0x3e6
testing.startAlarm.func1()
/usr/local/go/src/testing/testing.go:725 +0x14b
created by time.goFunc
/usr/local/go/src/time/sleep.go:129 +0x3a
goroutine 1 [chan receive]:
testing.RunTests(0x1d17878, 0x24a2540, 0x10f, 0x10f, 0xc820215b01)
/usr/local/go/src/testing/testing.go:583 +0x8d2
testing.(*M).Run(0xc820045ef8, 0xc820265c98)
/usr/local/go/src/testing/testing.go:515 +0x81
github.com/cockroachdb/cockroach/storage_test.TestMain(0xc820045ef8)
/go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de
main.main()
github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114
goroutine 17 [syscall, 1 minutes, locked to thread]:
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:1998 +0x1
goroutine 20 [chan receive]:
github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x277a0a0)
/go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64
created by github.com/cockroachdb/cockroach/util/log.init.1
/go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a
goroutine 14674 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4a0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14635 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef20)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14640 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9f040)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14669 [select]:
github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc820e94960)
/go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21
github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm()
/go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d46800)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14638 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efc0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14637 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef80)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14678 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6a0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14675 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14690 [select]:
github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1()
/go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc8215ff040)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14670 [select]:
github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820eaede0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14665 [semacquire]:
sync.runtime_Syncsemacquire(0xc821728f90)
/usr/local/go/src/runtime/sema.go:241 +0x201
sync.(*Cond).Wait(0xc821728f80)
/usr/local/go/src/sync/cond.go:63 +0x9b
github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc821800310)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb
github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821800310)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7
github.com/cockroachdb/cockroach/storage_test.TestTxnPutOutOfOrder(0xc8219150e0)
/go/src/github.com/cockroachdb/cockroach/storage/client_replica_test.go:322 +0xdaa
testing.tRunner(0xc8219150e0, 0x24a3cb0)
/usr/local/go/src/testing/testing.go:473 +0x98
created by testing.RunTests
/usr/local/go/src/testing/testing.go:582 +0x892
goroutine 14679 [select]:
github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc821801260, 0xecece9eac, 0xaee4eb8, 0x27794c0, 0xc821728f40, 0xc821800310, 0xc821926300, 0xc820021e98)
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568
github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc821926300, 0xc821224240)
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78
github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821224690, 0xc821729e00)
/go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246
github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14639 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efe0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14666 [chan receive]:
github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82185dda0)
/go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a
created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open
/go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4
goroutine 14636 [runnable]:
fmt.(*pp).argNumber(0xc821dd2ea0, 0x2, 0x1960640, 0xa, 0x8, 0x3, 0x1, 0x5, 0x0)
/usr/local/go/src/fmt/print.go:1088
fmt.(*pp).doPrintf(0xc821dd2ea0, 0x1960640, 0xa, 0xc820914048, 0x3, 0x3)
/usr/local/go/src/fmt/print.go:1144 +0xc87
fmt.Sprintf(0x1960640, 0xa, 0xc820914048, 0x3, 0x3, 0x0, 0x0)
/usr/local/go/src/fmt/print.go:203 +0x6f
github.com/cockroachdb/cockroach/roachpb.BatchRequest.String(0xc8, 0xdf388, 0x100000001, 0x1, 0x2, 0x0, 0x0, 0x0, 0xc8219d8ba0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/roachpb/batch.go:525 +0x64a
github.com/cockroachdb/cockroach/roachpb.(*BatchRequest).String(0xc820c99340, 0x0, 0x0)
<autogenerated>:414 +0xb3
--
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef60)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14673 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).Start.func4()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc480)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14671 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d47dc0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
FAIL github.com/cockroachdb/cockroach/storage 70.049s
=== RUN TestBatchBasics
I160518 17:34:06.434263 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
I160518 17:34:06.436714 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.02s)
=== RUN TestBatchRepr
I160518 17:34:06.441427 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
I160518 17:34:06.443054 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance
--- PASS: TestBatchRepr (0.02s)
=== RUN TestBatchGet
I160518 17:34:06.460742 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
```
Please assign, take a look and update the issue accordingly. | 1.0 | circleci: failed tests: TestTxnPutOutOfOrder - The following test appears to have failed:
[#18040](https://circleci.com/gh/cockroachdb/cockroach/18040):
```
I160518 17:35:26.740337 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740471 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740604 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740729 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
I160518 17:35:26.740859 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil>
panic: test timed out after 1m10s
goroutine 14692 [running]:
panic(0x156f960, 0xc821c80050)
/usr/local/go/src/runtime/panic.go:464 +0x3e6
testing.startAlarm.func1()
/usr/local/go/src/testing/testing.go:725 +0x14b
created by time.goFunc
/usr/local/go/src/time/sleep.go:129 +0x3a
goroutine 1 [chan receive]:
testing.RunTests(0x1d17878, 0x24a2540, 0x10f, 0x10f, 0xc820215b01)
/usr/local/go/src/testing/testing.go:583 +0x8d2
testing.(*M).Run(0xc820045ef8, 0xc820265c98)
/usr/local/go/src/testing/testing.go:515 +0x81
github.com/cockroachdb/cockroach/storage_test.TestMain(0xc820045ef8)
/go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de
main.main()
github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114
goroutine 17 [syscall, 1 minutes, locked to thread]:
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:1998 +0x1
goroutine 20 [chan receive]:
github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x277a0a0)
/go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64
created by github.com/cockroachdb/cockroach/util/log.init.1
/go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a
goroutine 14674 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4a0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14635 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef20)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14640 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9f040)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14669 [select]:
github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc820e94960)
/go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21
github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm()
/go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d46800)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14638 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efc0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14637 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef80)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14678 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6a0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14675 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14690 [select]:
github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1()
/go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc8215ff040)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14670 [select]:
github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820eaede0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14665 [semacquire]:
sync.runtime_Syncsemacquire(0xc821728f90)
/usr/local/go/src/runtime/sema.go:241 +0x201
sync.(*Cond).Wait(0xc821728f80)
/usr/local/go/src/sync/cond.go:63 +0x9b
github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc821800310)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb
github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821800310)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7
github.com/cockroachdb/cockroach/storage_test.TestTxnPutOutOfOrder(0xc8219150e0)
/go/src/github.com/cockroachdb/cockroach/storage/client_replica_test.go:322 +0xdaa
testing.tRunner(0xc8219150e0, 0x24a3cb0)
/usr/local/go/src/testing/testing.go:473 +0x98
created by testing.RunTests
/usr/local/go/src/testing/testing.go:582 +0x892
goroutine 14679 [select]:
github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc821801260, 0xecece9eac, 0xaee4eb8, 0x27794c0, 0xc821728f40, 0xc821800310, 0xc821926300, 0xc820021e98)
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568
github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc821926300, 0xc821224240)
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78
github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821224690, 0xc821729e00)
/go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246
github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14639 [select]:
github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efe0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14666 [chan receive]:
github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82185dda0)
/go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a
created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open
/go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4
goroutine 14636 [runnable]:
fmt.(*pp).argNumber(0xc821dd2ea0, 0x2, 0x1960640, 0xa, 0x8, 0x3, 0x1, 0x5, 0x0)
/usr/local/go/src/fmt/print.go:1088
fmt.(*pp).doPrintf(0xc821dd2ea0, 0x1960640, 0xa, 0xc820914048, 0x3, 0x3)
/usr/local/go/src/fmt/print.go:1144 +0xc87
fmt.Sprintf(0x1960640, 0xa, 0xc820914048, 0x3, 0x3, 0x0, 0x0)
/usr/local/go/src/fmt/print.go:203 +0x6f
github.com/cockroachdb/cockroach/roachpb.BatchRequest.String(0xc8, 0xdf388, 0x100000001, 0x1, 0x2, 0x0, 0x0, 0x0, 0xc8219d8ba0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/roachpb/batch.go:525 +0x64a
github.com/cockroachdb/cockroach/roachpb.(*BatchRequest).String(0xc820c99340, 0x0, 0x0)
<autogenerated>:414 +0xb3
--
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef60)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14673 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).Start.func4()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc480)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
goroutine 14671 [select]:
github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d47dc0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62
FAIL github.com/cockroachdb/cockroach/storage 70.049s
=== RUN TestBatchBasics
I160518 17:34:06.434263 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
I160518 17:34:06.436714 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.02s)
=== RUN TestBatchRepr
I160518 17:34:06.441427 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
I160518 17:34:06.443054 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance
--- PASS: TestBatchRepr (0.02s)
=== RUN TestBatchGet
I160518 17:34:06.460742 storage/engine/rocksdb.go:137 opening in memory rocksdb instance
```
Please assign, take a look and update the issue accordingly. | test | circleci failed tests testtxnputoutoforder the following test appears to have failed kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is panic test timed out after goroutine panic usr local go src runtime panic go testing startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach storage test testmain go src github com cockroachdb cockroach storage main test go main main github com cockroachdb cockroach storage test testmain go goroutine runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach util log loggingt flushdaemon go src github com cockroachdb cockroach util log clog go created by github com cockroachdb cockroach util log init go src github com cockroachdb cockroach util log clog go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach kv txncoordsender startstats go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach kv txncoordsender github com cockroachdb cockroach kv startstats fm go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage idallocator start go src github com cockroachdb cockroach storage id alloc go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage storepool start go src github com cockroachdb cockroach storage store pool go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine sync runtime syncsemacquire usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach util stop stopper quiesce go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach util stop stopper stop go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach storage test testtxnputoutoforder go src github com cockroachdb cockroach storage client replica test go testing trunner usr local go src testing testing go created by testing runtests usr local go src testing testing go goroutine github com cockroachdb cockroach storage replicascanner waitandprocess go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage storerangeset visit go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go created by github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go goroutine fmt pp argnumber usr local go src fmt print go fmt pp doprintf usr local go src fmt print go fmt sprintf usr local go src fmt print go github com cockroachdb cockroach roachpb batchrequest string go src github com cockroachdb cockroach roachpb batch go github com cockroachdb cockroach roachpb batchrequest string github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store start go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store processraft go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go fail github com cockroachdb cockroach storage run testbatchbasics storage engine rocksdb go opening in memory rocksdb instance storage engine rocksdb go closing in memory rocksdb instance pass testbatchbasics run testbatchrepr storage engine rocksdb go opening in memory rocksdb instance storage engine rocksdb go closing in memory rocksdb instance pass testbatchrepr run testbatchget storage engine rocksdb go opening in memory rocksdb instance please assign take a look and update the issue accordingly | 1 |
209,017 | 15,961,986,764 | IssuesEvent | 2021-04-16 00:12:37 | Carlosma7/TFM-GrandQuiz | https://api.github.com/repos/Carlosma7/TFM-GrandQuiz | closed | Definición tests clase Partida | test | Se deben definir los tests asociados a la clase Partida definida en #46 . | 1.0 | Definición tests clase Partida - Se deben definir los tests asociados a la clase Partida definida en #46 . | test | definición tests clase partida se deben definir los tests asociados a la clase partida definida en | 1 |
286,420 | 8,788,060,385 | IssuesEvent | 2018-12-20 20:48:58 | RPGHacker/asar | https://api.github.com/repos/RPGHacker/asar | closed | Segfault with defines and quotes and stuff | bug priority: medium | ```
!w = "lol""lol"
print "!w"
```
Segfaults.
The 2nd line gets parsed to `print "lol"lol"` after define evaluation, which causes qsplit to throw up. | 1.0 | Segfault with defines and quotes and stuff - ```
!w = "lol""lol"
print "!w"
```
Segfaults.
The 2nd line gets parsed to `print "lol"lol"` after define evaluation, which causes qsplit to throw up. | non_test | segfault with defines and quotes and stuff w lol lol print w segfaults the line gets parsed to print lol lol after define evaluation which causes qsplit to throw up | 0 |
136,566 | 18,741,071,817 | IssuesEvent | 2021-11-04 13:40:26 | ioana-github-enterprise/testremediate | https://api.github.com/repos/ioana-github-enterprise/testremediate | opened | CVE-2021-32769 (High) detected in micronaut-inject-2.3.1.jar, micronaut-core-2.3.1.jar | security vulnerability | ## CVE-2021-32769 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>micronaut-inject-2.3.1.jar</b>, <b>micronaut-core-2.3.1.jar</b></p></summary>
<p>
<details><summary><b>micronaut-inject-2.3.1.jar</b></p></summary>
<p>Natively Cloud Native</p>
<p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p>
<p>Path to vulnerable library: /micronaut-inject-2.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **micronaut-inject-2.3.1.jar** (Vulnerable Library)
</details>
<details><summary><b>micronaut-core-2.3.1.jar</b></p></summary>
<p>Natively Cloud Native</p>
<p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p>
<p>Path to vulnerable library: /micronaut-core-2.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **micronaut-core-2.3.1.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/testremediate/commit/c50f50f9e3a35342bd1f045ceb06f33a49e525fb">c50f50f9e3a35342bd1f045ceb06f33a49e525fb</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using "/../../" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.
<p>Publish Date: 2021-07-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769>CVE-2021-32769</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769</a></p>
<p>Release Date: 2021-07-16</p>
<p>Fix Resolution: io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-inject","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-inject:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"},{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-core","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-core:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32769","vulnerabilityDetails":"Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using \"/../../\" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-32769 (High) detected in micronaut-inject-2.3.1.jar, micronaut-core-2.3.1.jar - ## CVE-2021-32769 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>micronaut-inject-2.3.1.jar</b>, <b>micronaut-core-2.3.1.jar</b></p></summary>
<p>
<details><summary><b>micronaut-inject-2.3.1.jar</b></p></summary>
<p>Natively Cloud Native</p>
<p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p>
<p>Path to vulnerable library: /micronaut-inject-2.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **micronaut-inject-2.3.1.jar** (Vulnerable Library)
</details>
<details><summary><b>micronaut-core-2.3.1.jar</b></p></summary>
<p>Natively Cloud Native</p>
<p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p>
<p>Path to vulnerable library: /micronaut-core-2.3.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **micronaut-core-2.3.1.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/testremediate/commit/c50f50f9e3a35342bd1f045ceb06f33a49e525fb">c50f50f9e3a35342bd1f045ceb06f33a49e525fb</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using "/../../" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.
<p>Publish Date: 2021-07-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769>CVE-2021-32769</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769</a></p>
<p>Release Date: 2021-07-16</p>
<p>Fix Resolution: io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-inject","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-inject:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"},{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-core","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-core:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32769","vulnerabilityDetails":"Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using \"/../../\" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in micronaut inject jar micronaut core jar cve high severity vulnerability vulnerable libraries micronaut inject jar micronaut core jar micronaut inject jar natively cloud native library home page a href path to vulnerable library micronaut inject jar dependency hierarchy x micronaut inject jar vulnerable library micronaut core jar natively cloud native library home page a href path to vulnerable library micronaut core jar dependency hierarchy x micronaut core jar vulnerable library found in head commit a href found in base branch main vulnerability details micronaut is a jvm based full stack java framework designed for building jvm applications a path traversal vulnerability exists in versions prior to with a basic configuration it is possible to access any file from a filesystem using in the url this occurs because micronaut does not restrict file access to configured paths the vulnerability is patched in version as a workaround do not use in mapping use only which exposes only flat structure of a directory not allowing traversal if using linux another workaround is to run micronaut in chroot publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io micronaut micronaut core io micronaut micronaut inject isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree io micronaut micronaut inject isminimumfixversionavailable true minimumfixversion io micronaut micronaut core io micronaut micronaut inject packagetype java groupid io micronaut packagename micronaut core packageversion packagefilepaths istransitivedependency false dependencytree io micronaut micronaut core isminimumfixversionavailable true minimumfixversion io micronaut micronaut core io micronaut micronaut inject basebranches vulnerabilityidentifier cve vulnerabilitydetails micronaut is a jvm based full stack java framework designed for building jvm applications a path traversal vulnerability exists in versions prior to with a basic configuration it is possible to access any file from a filesystem using in the url this occurs because micronaut does not restrict file access to configured paths the vulnerability is patched in version as a workaround do not use in mapping use only which exposes only flat structure of a directory not allowing traversal if using linux another workaround is to run micronaut in chroot vulnerabilityurl | 0 |
85,653 | 3,697,239,299 | IssuesEvent | 2016-02-27 14:53:44 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Creating HPA using docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml fails validation | area/autoscaling kind/bug priority/P1 team/control-plane | I ran:
```
$ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml
```
which gave:
```
error validating "docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml": error validating data: found invalid field namespace for v1beta1.SubresourceReference; if you choose to ignore these errors, turn validation off with --validate=false
```
Adding validate=false ran fine.
```
$ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml --validate=false
```
I was running HEAD code locally.
cc @piosz @kubernetes/autoscaling | 1.0 | Creating HPA using docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml fails validation - I ran:
```
$ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml
```
which gave:
```
error validating "docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml": error validating data: found invalid field namespace for v1beta1.SubresourceReference; if you choose to ignore these errors, turn validation off with --validate=false
```
Adding validate=false ran fine.
```
$ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml --validate=false
```
I was running HEAD code locally.
cc @piosz @kubernetes/autoscaling | non_test | creating hpa using docs user guide horizontal pod autoscaling hpa php apache yaml fails validation i ran cluster kubectl sh create f docs user guide horizontal pod autoscaling hpa php apache yaml which gave error validating docs user guide horizontal pod autoscaling hpa php apache yaml error validating data found invalid field namespace for subresourcereference if you choose to ignore these errors turn validation off with validate false adding validate false ran fine cluster kubectl sh create f docs user guide horizontal pod autoscaling hpa php apache yaml validate false i was running head code locally cc piosz kubernetes autoscaling | 0 |
56,812 | 6,529,126,406 | IssuesEvent | 2017-08-30 10:16:01 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | Prevent employee update via employee request for DISMISSED employee | kind/bug priority/high status/test | Prevent employee update via employee request for DISMISSED employee
Return 409 with message "DISMISSED employee can't be updated"
**Request:**
```
curl -X POST \
http://dev.ehealth.world/api/employee_requests \
-H 'authorization: Bearer bVp1b1M4TGEzNWcvMWFMa1RmRW5iUT09' \
-H 'cache-control: no-cache' \
-H 'content-type: application/json' \
-H 'postman-token: d391baef-b66c-0fb3-c561-baec99a19f03' \
-d '{
"employee_request": {
"legal_entity_id": "f41ebb33-e9d0-4ef4-8e92-ee719d16554e",
"employee_id":"33afca77-9e1e-4c79-8a5a-fc8164b4da85",
"position": "P1",
"start_date": "2017-03-02",
"status": "NEW",
"employee_type": "DOCTOR",
"party": {
"first_name": "Петро",
"last_name": "Іванов",
"second_name": "Миколайович",
"birth_date": "1991-08-25",
"gender": "MALE",
"tax_id": "3067305998",
"email": "sp.virny+51@gmail.com",
"documents": [
{
"type": "PASSPORT",
"number": "120518"
}
],
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
]
},
"doctor": {
"educations": [
{
"country": "UA",
"city": "Київ",
"institution_name": "Академія Богомольця",
"issued_date": "2017-08-05",
"diploma_number": "DD123543",
"degree": "MASTER",
"speciality": "Педіатр"
}
],
"qualifications": [
{
"type": "CLINICAL_RESIDENCY",
"institution_name": "Академія Богомольця",
"speciality": "Педіатр",
"issued_date": "2017-08-05",
"certificate_number": "2017-08-05"
}
],
"specialities": [
{
"speciality": "PEDIATRICIAN",
"speciality_officio": true,
"level": "FIRST",
"qualification_type": "AWARDING",
"attestation_name": "Академія Богомольця",
"attestation_date": "2017-08-05",
"valid_to_date": "2017-08-05",
"certificate_number": "AB/21331"
}
],
"science_degree": {
"country": "UA",
"city": "Київ",
"degree": "PhD",
"institution_name": "Академія Богомольця",
"diploma_number": "DD123543",
"speciality": "THERAPIST",
"issued_date": "2017-08-05"
}
}
}
}'
```
**Actual result:**
200 Success
**Expected result:**
409 with message "DISMISSED employee can't be updated" | 1.0 | Prevent employee update via employee request for DISMISSED employee - Prevent employee update via employee request for DISMISSED employee
Return 409 with message "DISMISSED employee can't be updated"
**Request:**
```
curl -X POST \
http://dev.ehealth.world/api/employee_requests \
-H 'authorization: Bearer bVp1b1M4TGEzNWcvMWFMa1RmRW5iUT09' \
-H 'cache-control: no-cache' \
-H 'content-type: application/json' \
-H 'postman-token: d391baef-b66c-0fb3-c561-baec99a19f03' \
-d '{
"employee_request": {
"legal_entity_id": "f41ebb33-e9d0-4ef4-8e92-ee719d16554e",
"employee_id":"33afca77-9e1e-4c79-8a5a-fc8164b4da85",
"position": "P1",
"start_date": "2017-03-02",
"status": "NEW",
"employee_type": "DOCTOR",
"party": {
"first_name": "Петро",
"last_name": "Іванов",
"second_name": "Миколайович",
"birth_date": "1991-08-25",
"gender": "MALE",
"tax_id": "3067305998",
"email": "sp.virny+51@gmail.com",
"documents": [
{
"type": "PASSPORT",
"number": "120518"
}
],
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
]
},
"doctor": {
"educations": [
{
"country": "UA",
"city": "Київ",
"institution_name": "Академія Богомольця",
"issued_date": "2017-08-05",
"diploma_number": "DD123543",
"degree": "MASTER",
"speciality": "Педіатр"
}
],
"qualifications": [
{
"type": "CLINICAL_RESIDENCY",
"institution_name": "Академія Богомольця",
"speciality": "Педіатр",
"issued_date": "2017-08-05",
"certificate_number": "2017-08-05"
}
],
"specialities": [
{
"speciality": "PEDIATRICIAN",
"speciality_officio": true,
"level": "FIRST",
"qualification_type": "AWARDING",
"attestation_name": "Академія Богомольця",
"attestation_date": "2017-08-05",
"valid_to_date": "2017-08-05",
"certificate_number": "AB/21331"
}
],
"science_degree": {
"country": "UA",
"city": "Київ",
"degree": "PhD",
"institution_name": "Академія Богомольця",
"diploma_number": "DD123543",
"speciality": "THERAPIST",
"issued_date": "2017-08-05"
}
}
}
}'
```
**Actual result:**
200 Success
**Expected result:**
409 with message "DISMISSED employee can't be updated" | test | prevent employee update via employee request for dismissed employee prevent employee update via employee request for dismissed employee return with message dismissed employee can t be updated request curl x post h authorization bearer h cache control no cache h content type application json h postman token d employee request legal entity id employee id position start date status new employee type doctor party first name петро last name іванов second name миколайович birth date gender male tax id email sp virny gmail com documents type passport number phones type mobile number doctor educations country ua city київ institution name академія богомольця issued date diploma number degree master speciality педіатр qualifications type clinical residency institution name академія богомольця speciality педіатр issued date certificate number specialities speciality pediatrician speciality officio true level first qualification type awarding attestation name академія богомольця attestation date valid to date certificate number ab science degree country ua city київ degree phd institution name академія богомольця diploma number speciality therapist issued date actual result success expected result with message dismissed employee can t be updated | 1 |
591,359 | 17,837,887,088 | IssuesEvent | 2021-09-03 05:41:51 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | minikube start fails with podman or kvm2 drivers on s390x | kind/feature priority/backlog | minikube claims to run on s390x, but I have not had success yet. I am on Fedora 33 here, so perhaps another disto would would better, but I've seen Ubuntu fail similarly using the kvm2 driver.
<!--- Please include the "minikube start" command you used in your reproduction steps --->
**Steps to reproduce the issue:**
1. minikube config set driver podman && minikube delete
2. minikube start
*or*
1. minikube config set driver kvm2 && minikube delete
2. minikube start
**Full output of `minikube logs` command (for `podman`)**
<details>
==> Last Start <==
Log file created at: 2021/06/15 14:42:08
Running on machine: minikube1
Binary: Built with gc go1.16.4 for linux/s390x
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0615 14:42:08.728959 30552 out.go:291] Setting OutFile to fd 1 ...
I0615 14:42:08.729059 30552 out.go:343] isatty.IsTerminal(1) = true
I0615 14:42:08.729067 30552 out.go:304] Setting ErrFile to fd 2...
I0615 14:42:08.729071 30552 out.go:343] isatty.IsTerminal(2) = true
I0615 14:42:08.729176 30552 root.go:316] Updating PATH: /home/fedora/.minikube/bin
I0615 14:42:08.729372 30552 out.go:298] Setting JSON to false
I0615 14:42:08.730027 30552 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3644,"bootTime":1623764485,"procs":351,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"}
I0615 14:42:08.730074 30552 start.go:121] virtualization: kvm host
I0615 14:42:08.732896 30552 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x)
I0615 14:42:08.733072 30552 notify.go:169] Checking for updates...
I0615 14:42:08.733074 30552 driver.go:335] Setting default libvirt URI to qemu:///system
I0615 14:42:08.893533 30552 podman.go:121] podman version: 3.1.2
I0615 14:42:08.894674 30552 out.go:170] ✨ Using the podman driver based on user configuration
I0615 14:42:08.894726 30552 start.go:279] selected driver: podman
I0615 14:42:08.894729 30552 start.go:752] validating driver "podman" against <nil>
I0615 14:42:08.894736 30552 start.go:763] status for podman: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc:}
I0615 14:42:08.894791 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:09.043747 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1246912512 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 43.41s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:09.043798 30552 start_flags.go:259] no existing cluster config was found, will generate one from the flags
I0615 14:42:09.043940 30552 start_flags.go:311] Using suggested 2200MB memory alloc based on sys=4004MB, container=4004MB
I0615 14:42:09.044002 30552 start_flags.go:638] Wait components to verify : map[apiserver:true system_pods:true]
I0615 14:42:09.044013 30552 cni.go:93] Creating CNI manager for ""
I0615 14:42:09.044019 30552 cni.go:167] CNI unnecessary in this configuration, recommending no CNI
I0615 14:42:09.044022 30552 start_flags.go:273] config:
{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false}
I0615 14:42:09.045663 30552 out.go:170] 👍 Starting control plane node minikube in cluster minikube
I0615 14:42:09.045679 30552 cache.go:115] Beginning downloading kic base image for podman with docker
I0615 14:42:09.046691 30552 out.go:170] 🚜 Pulling base image ...
I0615 14:42:09.046719 30552 preload.go:110] Checking if preload exists for k8s version v1.20.7 and runtime docker
I0615 14:42:09.046784 30552 cache.go:134] Downloading gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache
I0615 14:42:09.046910 30552 image.go:58] Checking for gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 in local cache directory
I0615 14:42:09.046927 30552 image.go:118] Writing gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache
W0615 14:42:09.106869 30552 preload.go:140] https://storage.googleapis.com/minikube-preloaded-volume-tarballs/preloaded-images-k8s-v11-v1.20.7-docker-overlay2-s390x.tar.lz4 status code: 404
I0615 14:42:09.107061 30552 profile.go:148] Saving config to /home/fedora/.minikube/profiles/minikube/config.json ...
I0615 14:42:09.107073 30552 lock.go:36] WriteFile acquiring /home/fedora/.minikube/profiles/minikube/config.json: {Name:mkb8d756e7e807b96e4ac95488558156b33e3a18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0615 14:42:09.107211 30552 cache.go:108] acquiring lock: {Name:mk848f9056a0b7e8deba61c4d9ce1ec9447fb24f Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107243 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 exists
I0615 14:42:09.107250 30552 cache.go:97] cache image "docker.io/kubernetesui/metrics-scraper:v1.0.4" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4" took 40.935µs
I0615 14:42:09.107256 30552 cache.go:81] save to tar file docker.io/kubernetesui/metrics-scraper:v1.0.4 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 succeeded
I0615 14:42:09.107265 30552 cache.go:108] acquiring lock: {Name:mke06d93f1117b77439f99f030a503df9447235e Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107293 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 exists
I0615 14:42:09.107299 30552 cache.go:97] cache image "k8s.gcr.io/kube-apiserver:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7" took 35.361µs
I0615 14:42:09.107304 30552 cache.go:81] save to tar file k8s.gcr.io/kube-apiserver:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 succeeded
I0615 14:42:09.107312 30552 cache.go:108] acquiring lock: {Name:mk2f2a385b1cfc81a9ca9ce72bba6acd289ea36a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107339 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 exists
I0615 14:42:09.107344 30552 cache.go:97] cache image "k8s.gcr.io/kube-controller-manager:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7" took 33.518µs
I0615 14:42:09.107349 30552 cache.go:81] save to tar file k8s.gcr.io/kube-controller-manager:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 succeeded
I0615 14:42:09.107357 30552 cache.go:108] acquiring lock: {Name:mkbf90ca757cecf35c48d78ea90ab30ec4c88b47 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107383 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 exists
I0615 14:42:09.107388 30552 cache.go:97] cache image "k8s.gcr.io/kube-scheduler:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7" took 32.516µs
I0615 14:42:09.107393 30552 cache.go:81] save to tar file k8s.gcr.io/kube-scheduler:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 succeeded
I0615 14:42:09.107400 30552 cache.go:108] acquiring lock: {Name:mk42495b17f7ffd0cee7bc305c566b5bc1271eab Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107426 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 exists
I0615 14:42:09.107432 30552 cache.go:97] cache image "k8s.gcr.io/kube-proxy:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7" took 32.595µs
I0615 14:42:09.107437 30552 cache.go:81] save to tar file k8s.gcr.io/kube-proxy:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 succeeded
I0615 14:42:09.107445 30552 cache.go:108] acquiring lock: {Name:mk15a5e7d0f8daa4a4ee292d980fdb92a25be656 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107470 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 exists
I0615 14:42:09.107476 30552 cache.go:97] cache image "k8s.gcr.io/pause:3.2" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2" took 32.821µs
I0615 14:42:09.107480 30552 cache.go:81] save to tar file k8s.gcr.io/pause:3.2 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 succeeded
I0615 14:42:09.107488 30552 cache.go:108] acquiring lock: {Name:mke705c7f34a678272a870eb3ee742e676627306 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107513 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 exists
I0615 14:42:09.107519 30552 cache.go:97] cache image "k8s.gcr.io/etcd:3.4.13-0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0" took 32.046µs
I0615 14:42:09.107523 30552 cache.go:81] save to tar file k8s.gcr.io/etcd:3.4.13-0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 succeeded
I0615 14:42:09.107531 30552 cache.go:108] acquiring lock: {Name:mk58b9291a6bc97c1f1359b79ec84521b6f848fc Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107557 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 exists
I0615 14:42:09.107563 30552 cache.go:97] cache image "k8s.gcr.io/coredns:1.7.0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0" took 33.11µs
I0615 14:42:09.107567 30552 cache.go:81] save to tar file k8s.gcr.io/coredns:1.7.0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 succeeded
I0615 14:42:09.107575 30552 cache.go:108] acquiring lock: {Name:mk92ab5604f143a7cec05887d019b5678bdb0226 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107602 30552 cache.go:116] /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 exists
I0615 14:42:09.107607 30552 cache.go:97] cache image "gcr.io/k8s-minikube/storage-provisioner:v5" -> "/home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5" took 33.748µs
I0615 14:42:09.107613 30552 cache.go:81] save to tar file gcr.io/k8s-minikube/storage-provisioner:v5 -> /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 succeeded
I0615 14:42:09.107620 30552 cache.go:108] acquiring lock: {Name:mk9e35de248c46aa08df049e2c423012df463dc5 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107645 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 exists
I0615 14:42:09.107651 30552 cache.go:97] cache image "docker.io/kubernetesui/dashboard:v2.1.0" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0" took 31.152µs
I0615 14:42:09.107655 30552 cache.go:81] save to tar file docker.io/kubernetesui/dashboard:v2.1.0 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 succeeded
I0615 14:42:09.107658 30552 cache.go:88] Successfully saved all images to host disk.
E0615 14:42:09.283440 30552 cache.go:197] Error downloading kic artifacts: not yet implemented, see issue #8426
I0615 14:42:09.283450 30552 cache.go:202] Successfully downloaded all kic artifacts
I0615 14:42:09.283462 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.283495 30552 start.go:317] acquired machines lock for "minikube" in 27.196µs
I0615 14:42:09.283504 30552 start.go:89] Provisioning new machine with config: &{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} &{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}
I0615 14:42:09.283536 30552 start.go:126] createHost starting for "" (driver="podman")
I0615 14:42:09.285059 30552 out.go:197] 🔥 Creating podman container (CPUs=2, Memory=2200MB) ...
I0615 14:42:09.285218 30552 start.go:160] libmachine.API.Create for "minikube" (driver="podman")
I0615 14:42:09.285230 30552 client.go:168] LocalClient.Create starting
I0615 14:42:09.285267 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/ca.pem
I0615 14:42:09.285286 30552 main.go:128] libmachine: Decoding PEM data...
I0615 14:42:09.285297 30552 main.go:128] libmachine: Parsing certificate...
I0615 14:42:09.285368 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/cert.pem
I0615 14:42:09.285384 30552 main.go:128] libmachine: Decoding PEM data...
I0615 14:42:09.285393 30552 main.go:128] libmachine: Parsing certificate...
I0615 14:42:09.285627 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube --format "{{range .plugins}}{{if eq .type "bridge"}}{{(index (index .ipam.ranges 0) 0).subnet}},{{(index (index .ipam.ranges 0) 0).gateway}}{{end}}{{end}}"
I0615 14:42:09.393427 30552 network_create.go:67] Found existing network {name:minikube subnet:0xc000b75920 gateway:[0 0 0 0 0 0 0 0 0 0 255 255 192 168 49 1] mtu:0}
I0615 14:42:09.393438 30552 kic.go:106] calculated static IP "192.168.49.2" for the "minikube" container
I0615 14:42:09.393480 30552 cli_runner.go:115] Run: sudo -n podman ps -a --format {{.Names}}
I0615 14:42:09.523500 30552 cli_runner.go:115] Run: sudo -n podman volume create minikube --label name.minikube.sigs.k8s.io=minikube --label created_by.minikube.sigs.k8s.io=true
I0615 14:42:09.653396 30552 oci.go:102] Successfully created a podman volume minikube
I0615 14:42:09.653434 30552 cli_runner.go:115] Run: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib
W0615 14:42:10.133503 30552 cli_runner.go:162] sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib returned with exit code 125
I0615 14:42:10.133523 30552 client.go:171] LocalClient.Create took 848.289535ms
I0615 14:42:12.134090 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0615 14:42:12.134129 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:12.243575 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:12.353526 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:12.353575 30552 retry.go:31] will retry after 276.165072ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:12.630208 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:12.733566 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:12.843522 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:12.843573 30552 retry.go:31] will retry after 540.190908ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:13.384633 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:13.513535 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:13.623653 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:13.623710 30552 retry.go:31] will retry after 655.06503ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:14.279839 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:14.423423 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:14.543373 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
W0615 14:42:14.543426 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:14.543432 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:14.543436 30552 start.go:129] duration metric: createHost completed in 5.259896732s
I0615 14:42:14.543440 30552 start.go:80] releasing machines lock for "minikube", held for 5.259942055s
W0615 14:42:14.543449 30552 start.go:518] error starting host: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125
stdout:
stderr:
Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23...
no image found in manifest list for architecture s390x, variant "", OS linux
Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux
I0615 14:42:14.543773 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}}
W0615 14:42:14.683483 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125
I0615 14:42:14.683504 30552 delete.go:82] Unable to get host status for minikube, assuming it has already been deleted: state: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:14.683568 30552 out.go:235] 🤦 StartHost failed, but will try again: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125
stdout:
stderr:
Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23...
no image found in manifest list for architecture s390x, variant "", OS linux
Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux
I0615 14:42:14.683719 30552 start.go:533] Will try again in 5 seconds ...
I0615 14:42:19.684326 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:19.684380 30552 start.go:317] acquired machines lock for "minikube" in 42.164µs
I0615 14:42:19.684389 30552 start.go:93] Skipping create...Using existing machine configuration
I0615 14:42:19.684393 30552 fix.go:55] fixHost starting:
I0615 14:42:19.684563 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}}
W0615 14:42:19.813413 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125
I0615 14:42:19.813429 30552 fix.go:108] recreateIfNeeded on minikube: state= err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:19.813439 30552 fix.go:113] machineExists: true. err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:19.813444 30552 fix.go:134] unexpected machine state, will restart: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:19.815104 30552 out.go:170] 🔄 Restarting existing podman container for "minikube" ...
I0615 14:42:19.815145 30552 cli_runner.go:115] Run: sudo -n podman start --cgroup-manager cgroupfs minikube
W0615 14:42:19.953379 30552 cli_runner.go:162] sudo -n podman start --cgroup-manager cgroupfs minikube returned with exit code 125
I0615 14:42:19.953414 30552 cli_runner.go:115] Run: sudo -n podman inspect minikube
I0615 14:42:20.073749 30552 errors.go:84] Postmortem inspect ("sudo -n podman inspect minikube"): -- stdout --
[
{
"Name": "minikube",
"Driver": "local",
"Mountpoint": "/var/lib/containers/storage/volumes/minikube/_data",
"CreatedAt": "2021-06-15T14:42:09.614053773Z",
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"name.minikube.sigs.k8s.io": "minikube"
},
"Scope": "local",
"Options": {}
}
]
-- /stdout --
I0615 14:42:20.073838 30552 cli_runner.go:115] Run: sudo -n podman logs --timestamps minikube
W0615 14:42:20.203437 30552 cli_runner.go:162] sudo -n podman logs --timestamps minikube returned with exit code 125
W0615 14:42:20.203447 30552 errors.go:89] Failed to get postmortem logs. sudo -n podman logs --timestamps minikube :sudo -n podman logs --timestamps minikube: exit status 125
stdout:
stderr:
Error: no container with name or ID "minikube" found: no such container
I0615 14:42:20.203478 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:20.353543 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.353562 30552 errors.go:106] postmortem podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.353590 30552 network_create.go:255] running [podman network inspect minikube] to gather additional debugging logs...
I0615 14:42:20.353609 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube
I0615 14:42:20.483538 30552 network_create.go:260] output of [sudo -n podman network inspect minikube]: -- stdout --
[
{
"cniVersion": "0.4.0",
"name": "minikube",
"plugins": [
{
"bridge": "cni-podman1",
"hairpinMode": true,
"ipMasq": true,
"ipam": {
"ranges": [
[
{
"gateway": "192.168.49.1",
"subnet": "192.168.49.0/24"
}
]
],
"routes": [
{
"dst": "0.0.0.0/0"
}
],
"type": "host-local"
},
"isGateway": true,
"type": "bridge"
},
{
"capabilities": {
"portMappings": true
},
"type": "portmap"
},
{
"backend": "",
"type": "firewall"
},
{
"type": "tuning"
},
{
"capabilities": {
"aliases": true
},
"domainName": "dns.podman",
"type": "dnsname"
}
]
}
]
-- /stdout --
I0615 14:42:20.483583 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:20.643548 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243791360 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 55.01s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.643777 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube
W0615 14:42:20.793518 30552 cli_runner.go:162] sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube returned with exit code 125
I0615 14:42:20.793559 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0615 14:42:20.793592 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:20.943495 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:21.083399 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:21.083457 30552 retry.go:31] will retry after 234.428547ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:21.318877 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:21.443589 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:21.573394 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:21.573444 30552 retry.go:31] will retry after 346.739061ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:21.920931 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:22.053534 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:22.203360 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:22.203409 30552 retry.go:31] will retry after 667.587979ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:22.871144 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:22.993505 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:23.123357 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
W0615 14:42:23.123417 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:23.123423 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:23.123428 30552 fix.go:57] fixHost completed within 3.439034957s
I0615 14:42:23.123432 30552 start.go:80] releasing machines lock for "minikube", held for 3.43904856s
W0615 14:42:23.123519 30552 out.go:235] 😿 Failed to start podman container. Running "minikube delete" may fix it: podman inspect ip minikube: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:23.125747 30552 out.go:170]
W0615 14:42:23.125806 30552 out.go:235] ❌ Exiting due to GUEST_PROVISION: Failed to start host: podman inspect ip minikube: sudo -n podman container inspect -f minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:23.125913 30552 out.go:235]
W0615 14:42:23.126777 30552 out.go:235] ╭────────────────────────────────────────────────────────────────────╮
W0615 14:42:23.126788 30552 out.go:235] │ │
W0615 14:42:23.126807 30552 out.go:235] │ 😿 If the above advice does not help, please let us know: │
W0615 14:42:23.126827 30552 out.go:235] │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │
W0615 14:42:23.126842 30552 out.go:235] │ │
W0615 14:42:23.126856 30552 out.go:235] │ Please attach the following file to the GitHub issue: │
W0615 14:42:23.126867 30552 out.go:235] │ - /home/fedora/.minikube/logs/lastStart.txt │
W0615 14:42:23.126877 30552 out.go:235] │ │
W0615 14:42:23.126888 30552 out.go:235] ╰────────────────────────────────────────────────────────────────────╯
W0615 14:42:23.126900 30552 out.go:235]
❌ Exiting due to GUEST_STATUS: state: unknown state "minikube": sudo -n podman container inspect minikube --format=: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
╭───────────────────────────────────────────────────────────────────────────╮
│ │
│ 😿 If the above advice does not help, please let us know: │
│ 👉 https://github.com/kubernetes/minikube/issues/new/choose │
│ │
│ Please attach the following file to the GitHub issue: │
│ - /tmp/minikube_logs_f30b94c7b8be27a1785d74f9772c624a74c09c39_0.log │
│ │
╰───────────────────────────────────────────────────────────────────────────╯
</details>
**Full output of `minikube logs` command (for `kvm2`)**
<details>
==> Audit <==
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:05 UTC | Tue, 15 Jun 2021 13:57:05 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:13 UTC | Tue, 15 Jun 2021 13:57:13 UTC |
| start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:38 UTC | Tue, 15 Jun 2021 14:29:38 UTC |
| start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:49 UTC | Tue, 15 Jun 2021 14:29:49 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:35 UTC | Tue, 15 Jun 2021 14:37:35 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:40 UTC | Tue, 15 Jun 2021 14:37:40 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:18 UTC | Tue, 15 Jun 2021 14:40:18 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC |
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:06 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:07 UTC |
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC |
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
==> Last Start <==
Log file created at: 2021/06/15 14:45:09
Running on machine: minikube1
Binary: Built with gc go1.16.4 for linux/s390x
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0615 14:45:09.494355 35758 out.go:291] Setting OutFile to fd 1 ...
I0615 14:45:09.494462 35758 out.go:343] isatty.IsTerminal(1) = true
I0615 14:45:09.494465 35758 out.go:304] Setting ErrFile to fd 2...
I0615 14:45:09.494468 35758 out.go:343] isatty.IsTerminal(2) = true
I0615 14:45:09.494545 35758 root.go:316] Updating PATH: /home/fedora/.minikube/bin
I0615 14:45:09.494725 35758 out.go:298] Setting JSON to false
I0615 14:45:09.495146 35758 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3824,"bootTime":1623764485,"procs":92,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"}
I0615 14:45:09.495190 35758 start.go:121] virtualization: kvm host
I0615 14:45:09.496967 35758 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x)
I0615 14:45:09.497089 35758 notify.go:169] Checking for updates...
I0615 14:45:09.497486 35758 driver.go:335] Setting default libvirt URI to qemu:///system
I0615 14:45:09.498712 35758 out.go:170] ✨ Using the kvm2 driver based on user configuration
I0615 14:45:09.498722 35758 start.go:279] selected driver: kvm2
I0615 14:45:09.498725 35758 start.go:752] validating driver "kvm2" against <nil>
I0615 14:45:09.498731 35758 start.go:763] status for kvm2: {Installed:true Healthy:false Running:true NeedsImprovement:false Error:libvirt group membership check failed:
error getting current user's GIDs: user: GroupIds requires cgo Reason:PR_KVM_USER_PERMISSION Fix:Check that libvirtd is properly installed and that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) Doc:https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/}
I0615 14:45:09.499768 35758 out.go:170]
W0615 14:45:09.499816 35758 out.go:235] 🚫 Exiting due to PR_KVM_USER_PERMISSION: libvirt group membership check failed:
error getting current user's GIDs: user: GroupIds requires cgo
W0615 14:45:09.499939 35758 out.go:235] 💡 Suggestion: Ensure that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!)
W0615 14:45:09.499974 35758 out.go:235] 📘 Documentation: https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/
W0615 14:45:09.499985 35758 out.go:235] 🍿 Related issues:
W0615 14:45:09.500018 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/5617
W0615 14:45:09.500049 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/10070
🤷 Profile "minikube" not found. Run "minikube profile list" to view all profiles.
👉 To start a cluster, run: "minikube start"
</details>
| 1.0 | minikube start fails with podman or kvm2 drivers on s390x - minikube claims to run on s390x, but I have not had success yet. I am on Fedora 33 here, so perhaps another disto would would better, but I've seen Ubuntu fail similarly using the kvm2 driver.
<!--- Please include the "minikube start" command you used in your reproduction steps --->
**Steps to reproduce the issue:**
1. minikube config set driver podman && minikube delete
2. minikube start
*or*
1. minikube config set driver kvm2 && minikube delete
2. minikube start
**Full output of `minikube logs` command (for `podman`)**
<details>
==> Last Start <==
Log file created at: 2021/06/15 14:42:08
Running on machine: minikube1
Binary: Built with gc go1.16.4 for linux/s390x
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0615 14:42:08.728959 30552 out.go:291] Setting OutFile to fd 1 ...
I0615 14:42:08.729059 30552 out.go:343] isatty.IsTerminal(1) = true
I0615 14:42:08.729067 30552 out.go:304] Setting ErrFile to fd 2...
I0615 14:42:08.729071 30552 out.go:343] isatty.IsTerminal(2) = true
I0615 14:42:08.729176 30552 root.go:316] Updating PATH: /home/fedora/.minikube/bin
I0615 14:42:08.729372 30552 out.go:298] Setting JSON to false
I0615 14:42:08.730027 30552 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3644,"bootTime":1623764485,"procs":351,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"}
I0615 14:42:08.730074 30552 start.go:121] virtualization: kvm host
I0615 14:42:08.732896 30552 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x)
I0615 14:42:08.733072 30552 notify.go:169] Checking for updates...
I0615 14:42:08.733074 30552 driver.go:335] Setting default libvirt URI to qemu:///system
I0615 14:42:08.893533 30552 podman.go:121] podman version: 3.1.2
I0615 14:42:08.894674 30552 out.go:170] ✨ Using the podman driver based on user configuration
I0615 14:42:08.894726 30552 start.go:279] selected driver: podman
I0615 14:42:08.894729 30552 start.go:752] validating driver "podman" against <nil>
I0615 14:42:08.894736 30552 start.go:763] status for podman: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc:}
I0615 14:42:08.894791 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:09.043747 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1246912512 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 43.41s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:09.043798 30552 start_flags.go:259] no existing cluster config was found, will generate one from the flags
I0615 14:42:09.043940 30552 start_flags.go:311] Using suggested 2200MB memory alloc based on sys=4004MB, container=4004MB
I0615 14:42:09.044002 30552 start_flags.go:638] Wait components to verify : map[apiserver:true system_pods:true]
I0615 14:42:09.044013 30552 cni.go:93] Creating CNI manager for ""
I0615 14:42:09.044019 30552 cni.go:167] CNI unnecessary in this configuration, recommending no CNI
I0615 14:42:09.044022 30552 start_flags.go:273] config:
{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false}
I0615 14:42:09.045663 30552 out.go:170] 👍 Starting control plane node minikube in cluster minikube
I0615 14:42:09.045679 30552 cache.go:115] Beginning downloading kic base image for podman with docker
I0615 14:42:09.046691 30552 out.go:170] 🚜 Pulling base image ...
I0615 14:42:09.046719 30552 preload.go:110] Checking if preload exists for k8s version v1.20.7 and runtime docker
I0615 14:42:09.046784 30552 cache.go:134] Downloading gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache
I0615 14:42:09.046910 30552 image.go:58] Checking for gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 in local cache directory
I0615 14:42:09.046927 30552 image.go:118] Writing gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache
W0615 14:42:09.106869 30552 preload.go:140] https://storage.googleapis.com/minikube-preloaded-volume-tarballs/preloaded-images-k8s-v11-v1.20.7-docker-overlay2-s390x.tar.lz4 status code: 404
I0615 14:42:09.107061 30552 profile.go:148] Saving config to /home/fedora/.minikube/profiles/minikube/config.json ...
I0615 14:42:09.107073 30552 lock.go:36] WriteFile acquiring /home/fedora/.minikube/profiles/minikube/config.json: {Name:mkb8d756e7e807b96e4ac95488558156b33e3a18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0615 14:42:09.107211 30552 cache.go:108] acquiring lock: {Name:mk848f9056a0b7e8deba61c4d9ce1ec9447fb24f Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107243 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 exists
I0615 14:42:09.107250 30552 cache.go:97] cache image "docker.io/kubernetesui/metrics-scraper:v1.0.4" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4" took 40.935µs
I0615 14:42:09.107256 30552 cache.go:81] save to tar file docker.io/kubernetesui/metrics-scraper:v1.0.4 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 succeeded
I0615 14:42:09.107265 30552 cache.go:108] acquiring lock: {Name:mke06d93f1117b77439f99f030a503df9447235e Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107293 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 exists
I0615 14:42:09.107299 30552 cache.go:97] cache image "k8s.gcr.io/kube-apiserver:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7" took 35.361µs
I0615 14:42:09.107304 30552 cache.go:81] save to tar file k8s.gcr.io/kube-apiserver:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 succeeded
I0615 14:42:09.107312 30552 cache.go:108] acquiring lock: {Name:mk2f2a385b1cfc81a9ca9ce72bba6acd289ea36a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107339 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 exists
I0615 14:42:09.107344 30552 cache.go:97] cache image "k8s.gcr.io/kube-controller-manager:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7" took 33.518µs
I0615 14:42:09.107349 30552 cache.go:81] save to tar file k8s.gcr.io/kube-controller-manager:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 succeeded
I0615 14:42:09.107357 30552 cache.go:108] acquiring lock: {Name:mkbf90ca757cecf35c48d78ea90ab30ec4c88b47 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107383 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 exists
I0615 14:42:09.107388 30552 cache.go:97] cache image "k8s.gcr.io/kube-scheduler:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7" took 32.516µs
I0615 14:42:09.107393 30552 cache.go:81] save to tar file k8s.gcr.io/kube-scheduler:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 succeeded
I0615 14:42:09.107400 30552 cache.go:108] acquiring lock: {Name:mk42495b17f7ffd0cee7bc305c566b5bc1271eab Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107426 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 exists
I0615 14:42:09.107432 30552 cache.go:97] cache image "k8s.gcr.io/kube-proxy:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7" took 32.595µs
I0615 14:42:09.107437 30552 cache.go:81] save to tar file k8s.gcr.io/kube-proxy:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 succeeded
I0615 14:42:09.107445 30552 cache.go:108] acquiring lock: {Name:mk15a5e7d0f8daa4a4ee292d980fdb92a25be656 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107470 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 exists
I0615 14:42:09.107476 30552 cache.go:97] cache image "k8s.gcr.io/pause:3.2" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2" took 32.821µs
I0615 14:42:09.107480 30552 cache.go:81] save to tar file k8s.gcr.io/pause:3.2 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 succeeded
I0615 14:42:09.107488 30552 cache.go:108] acquiring lock: {Name:mke705c7f34a678272a870eb3ee742e676627306 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107513 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 exists
I0615 14:42:09.107519 30552 cache.go:97] cache image "k8s.gcr.io/etcd:3.4.13-0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0" took 32.046µs
I0615 14:42:09.107523 30552 cache.go:81] save to tar file k8s.gcr.io/etcd:3.4.13-0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 succeeded
I0615 14:42:09.107531 30552 cache.go:108] acquiring lock: {Name:mk58b9291a6bc97c1f1359b79ec84521b6f848fc Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107557 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 exists
I0615 14:42:09.107563 30552 cache.go:97] cache image "k8s.gcr.io/coredns:1.7.0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0" took 33.11µs
I0615 14:42:09.107567 30552 cache.go:81] save to tar file k8s.gcr.io/coredns:1.7.0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 succeeded
I0615 14:42:09.107575 30552 cache.go:108] acquiring lock: {Name:mk92ab5604f143a7cec05887d019b5678bdb0226 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107602 30552 cache.go:116] /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 exists
I0615 14:42:09.107607 30552 cache.go:97] cache image "gcr.io/k8s-minikube/storage-provisioner:v5" -> "/home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5" took 33.748µs
I0615 14:42:09.107613 30552 cache.go:81] save to tar file gcr.io/k8s-minikube/storage-provisioner:v5 -> /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 succeeded
I0615 14:42:09.107620 30552 cache.go:108] acquiring lock: {Name:mk9e35de248c46aa08df049e2c423012df463dc5 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.107645 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 exists
I0615 14:42:09.107651 30552 cache.go:97] cache image "docker.io/kubernetesui/dashboard:v2.1.0" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0" took 31.152µs
I0615 14:42:09.107655 30552 cache.go:81] save to tar file docker.io/kubernetesui/dashboard:v2.1.0 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 succeeded
I0615 14:42:09.107658 30552 cache.go:88] Successfully saved all images to host disk.
E0615 14:42:09.283440 30552 cache.go:197] Error downloading kic artifacts: not yet implemented, see issue #8426
I0615 14:42:09.283450 30552 cache.go:202] Successfully downloaded all kic artifacts
I0615 14:42:09.283462 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:09.283495 30552 start.go:317] acquired machines lock for "minikube" in 27.196µs
I0615 14:42:09.283504 30552 start.go:89] Provisioning new machine with config: &{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} &{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}
I0615 14:42:09.283536 30552 start.go:126] createHost starting for "" (driver="podman")
I0615 14:42:09.285059 30552 out.go:197] 🔥 Creating podman container (CPUs=2, Memory=2200MB) ...
I0615 14:42:09.285218 30552 start.go:160] libmachine.API.Create for "minikube" (driver="podman")
I0615 14:42:09.285230 30552 client.go:168] LocalClient.Create starting
I0615 14:42:09.285267 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/ca.pem
I0615 14:42:09.285286 30552 main.go:128] libmachine: Decoding PEM data...
I0615 14:42:09.285297 30552 main.go:128] libmachine: Parsing certificate...
I0615 14:42:09.285368 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/cert.pem
I0615 14:42:09.285384 30552 main.go:128] libmachine: Decoding PEM data...
I0615 14:42:09.285393 30552 main.go:128] libmachine: Parsing certificate...
I0615 14:42:09.285627 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube --format "{{range .plugins}}{{if eq .type "bridge"}}{{(index (index .ipam.ranges 0) 0).subnet}},{{(index (index .ipam.ranges 0) 0).gateway}}{{end}}{{end}}"
I0615 14:42:09.393427 30552 network_create.go:67] Found existing network {name:minikube subnet:0xc000b75920 gateway:[0 0 0 0 0 0 0 0 0 0 255 255 192 168 49 1] mtu:0}
I0615 14:42:09.393438 30552 kic.go:106] calculated static IP "192.168.49.2" for the "minikube" container
I0615 14:42:09.393480 30552 cli_runner.go:115] Run: sudo -n podman ps -a --format {{.Names}}
I0615 14:42:09.523500 30552 cli_runner.go:115] Run: sudo -n podman volume create minikube --label name.minikube.sigs.k8s.io=minikube --label created_by.minikube.sigs.k8s.io=true
I0615 14:42:09.653396 30552 oci.go:102] Successfully created a podman volume minikube
I0615 14:42:09.653434 30552 cli_runner.go:115] Run: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib
W0615 14:42:10.133503 30552 cli_runner.go:162] sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib returned with exit code 125
I0615 14:42:10.133523 30552 client.go:171] LocalClient.Create took 848.289535ms
I0615 14:42:12.134090 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0615 14:42:12.134129 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:12.243575 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:12.353526 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:12.353575 30552 retry.go:31] will retry after 276.165072ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:12.630208 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:12.733566 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:12.843522 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:12.843573 30552 retry.go:31] will retry after 540.190908ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:13.384633 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:13.513535 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:13.623653 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:13.623710 30552 retry.go:31] will retry after 655.06503ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:14.279839 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:14.423423 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:14.543373 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
W0615 14:42:14.543426 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:14.543432 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:14.543436 30552 start.go:129] duration metric: createHost completed in 5.259896732s
I0615 14:42:14.543440 30552 start.go:80] releasing machines lock for "minikube", held for 5.259942055s
W0615 14:42:14.543449 30552 start.go:518] error starting host: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125
stdout:
stderr:
Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23...
no image found in manifest list for architecture s390x, variant "", OS linux
Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux
I0615 14:42:14.543773 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}}
W0615 14:42:14.683483 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125
I0615 14:42:14.683504 30552 delete.go:82] Unable to get host status for minikube, assuming it has already been deleted: state: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:14.683568 30552 out.go:235] 🤦 StartHost failed, but will try again: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125
stdout:
stderr:
Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23...
no image found in manifest list for architecture s390x, variant "", OS linux
Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux
I0615 14:42:14.683719 30552 start.go:533] Will try again in 5 seconds ...
I0615 14:42:19.684326 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0615 14:42:19.684380 30552 start.go:317] acquired machines lock for "minikube" in 42.164µs
I0615 14:42:19.684389 30552 start.go:93] Skipping create...Using existing machine configuration
I0615 14:42:19.684393 30552 fix.go:55] fixHost starting:
I0615 14:42:19.684563 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}}
W0615 14:42:19.813413 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125
I0615 14:42:19.813429 30552 fix.go:108] recreateIfNeeded on minikube: state= err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:19.813439 30552 fix.go:113] machineExists: true. err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:19.813444 30552 fix.go:134] unexpected machine state, will restart: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:19.815104 30552 out.go:170] 🔄 Restarting existing podman container for "minikube" ...
I0615 14:42:19.815145 30552 cli_runner.go:115] Run: sudo -n podman start --cgroup-manager cgroupfs minikube
W0615 14:42:19.953379 30552 cli_runner.go:162] sudo -n podman start --cgroup-manager cgroupfs minikube returned with exit code 125
I0615 14:42:19.953414 30552 cli_runner.go:115] Run: sudo -n podman inspect minikube
I0615 14:42:20.073749 30552 errors.go:84] Postmortem inspect ("sudo -n podman inspect minikube"): -- stdout --
[
{
"Name": "minikube",
"Driver": "local",
"Mountpoint": "/var/lib/containers/storage/volumes/minikube/_data",
"CreatedAt": "2021-06-15T14:42:09.614053773Z",
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"name.minikube.sigs.k8s.io": "minikube"
},
"Scope": "local",
"Options": {}
}
]
-- /stdout --
I0615 14:42:20.073838 30552 cli_runner.go:115] Run: sudo -n podman logs --timestamps minikube
W0615 14:42:20.203437 30552 cli_runner.go:162] sudo -n podman logs --timestamps minikube returned with exit code 125
W0615 14:42:20.203447 30552 errors.go:89] Failed to get postmortem logs. sudo -n podman logs --timestamps minikube :sudo -n podman logs --timestamps minikube: exit status 125
stdout:
stderr:
Error: no container with name or ID "minikube" found: no such container
I0615 14:42:20.203478 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:20.353543 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.353562 30552 errors.go:106] postmortem podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.353590 30552 network_create.go:255] running [podman network inspect minikube] to gather additional debugging logs...
I0615 14:42:20.353609 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube
I0615 14:42:20.483538 30552 network_create.go:260] output of [sudo -n podman network inspect minikube]: -- stdout --
[
{
"cniVersion": "0.4.0",
"name": "minikube",
"plugins": [
{
"bridge": "cni-podman1",
"hairpinMode": true,
"ipMasq": true,
"ipam": {
"ranges": [
[
{
"gateway": "192.168.49.1",
"subnet": "192.168.49.0/24"
}
]
],
"routes": [
{
"dst": "0.0.0.0/0"
}
],
"type": "host-local"
},
"isGateway": true,
"type": "bridge"
},
{
"capabilities": {
"portMappings": true
},
"type": "portmap"
},
{
"backend": "",
"type": "firewall"
},
{
"type": "tuning"
},
{
"capabilities": {
"aliases": true
},
"domainName": "dns.podman",
"type": "dnsname"
}
]
}
]
-- /stdout --
I0615 14:42:20.483583 30552 cli_runner.go:115] Run: sudo -n podman system info --format json
I0615 14:42:20.643548 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243791360 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1
commit: 1535fedf0b83fb898d449f9680000f729ba719f5
spec: 1.0.0
+SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 55.01s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}}
I0615 14:42:20.643777 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube
W0615 14:42:20.793518 30552 cli_runner.go:162] sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube returned with exit code 125
I0615 14:42:20.793559 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0615 14:42:20.793592 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:20.943495 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:21.083399 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:21.083457 30552 retry.go:31] will retry after 234.428547ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:21.318877 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:21.443589 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:21.573394 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:21.573444 30552 retry.go:31] will retry after 346.739061ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:21.920931 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:22.053534 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:22.203360 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
I0615 14:42:22.203409 30552 retry.go:31] will retry after 667.587979ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:22.871144 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}}
I0615 14:42:22.993505 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube
W0615 14:42:23.123357 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125
W0615 14:42:23.123417 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:23.123423 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:23.123428 30552 fix.go:57] fixHost completed within 3.439034957s
I0615 14:42:23.123432 30552 start.go:80] releasing machines lock for "minikube", held for 3.43904856s
W0615 14:42:23.123519 30552 out.go:235] 😿 Failed to start podman container. Running "minikube delete" may fix it: podman inspect ip minikube: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
I0615 14:42:23.125747 30552 out.go:170]
W0615 14:42:23.125806 30552 out.go:235] ❌ Exiting due to GUEST_PROVISION: Failed to start host: podman inspect ip minikube: sudo -n podman container inspect -f minikube: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
W0615 14:42:23.125913 30552 out.go:235]
W0615 14:42:23.126777 30552 out.go:235] ╭────────────────────────────────────────────────────────────────────╮
W0615 14:42:23.126788 30552 out.go:235] │ │
W0615 14:42:23.126807 30552 out.go:235] │ 😿 If the above advice does not help, please let us know: │
W0615 14:42:23.126827 30552 out.go:235] │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │
W0615 14:42:23.126842 30552 out.go:235] │ │
W0615 14:42:23.126856 30552 out.go:235] │ Please attach the following file to the GitHub issue: │
W0615 14:42:23.126867 30552 out.go:235] │ - /home/fedora/.minikube/logs/lastStart.txt │
W0615 14:42:23.126877 30552 out.go:235] │ │
W0615 14:42:23.126888 30552 out.go:235] ╰────────────────────────────────────────────────────────────────────╯
W0615 14:42:23.126900 30552 out.go:235]
❌ Exiting due to GUEST_STATUS: state: unknown state "minikube": sudo -n podman container inspect minikube --format=: exit status 125
stdout:
stderr:
Error: error inspecting object: no such container minikube
╭───────────────────────────────────────────────────────────────────────────╮
│ │
│ 😿 If the above advice does not help, please let us know: │
│ 👉 https://github.com/kubernetes/minikube/issues/new/choose │
│ │
│ Please attach the following file to the GitHub issue: │
│ - /tmp/minikube_logs_f30b94c7b8be27a1785d74f9772c624a74c09c39_0.log │
│ │
╰───────────────────────────────────────────────────────────────────────────╯
</details>
**Full output of `minikube logs` command (for `kvm2`)**
<details>
==> Audit <==
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:05 UTC | Tue, 15 Jun 2021 13:57:05 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:13 UTC | Tue, 15 Jun 2021 13:57:13 UTC |
| start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:38 UTC | Tue, 15 Jun 2021 14:29:38 UTC |
| start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:49 UTC | Tue, 15 Jun 2021 14:29:49 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:35 UTC | Tue, 15 Jun 2021 14:37:35 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:40 UTC | Tue, 15 Jun 2021 14:37:40 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:18 UTC | Tue, 15 Jun 2021 14:40:18 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC |
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC |
| config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:06 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:07 UTC |
| config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC |
| delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC |
|---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------|
==> Last Start <==
Log file created at: 2021/06/15 14:45:09
Running on machine: minikube1
Binary: Built with gc go1.16.4 for linux/s390x
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0615 14:45:09.494355 35758 out.go:291] Setting OutFile to fd 1 ...
I0615 14:45:09.494462 35758 out.go:343] isatty.IsTerminal(1) = true
I0615 14:45:09.494465 35758 out.go:304] Setting ErrFile to fd 2...
I0615 14:45:09.494468 35758 out.go:343] isatty.IsTerminal(2) = true
I0615 14:45:09.494545 35758 root.go:316] Updating PATH: /home/fedora/.minikube/bin
I0615 14:45:09.494725 35758 out.go:298] Setting JSON to false
I0615 14:45:09.495146 35758 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3824,"bootTime":1623764485,"procs":92,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"}
I0615 14:45:09.495190 35758 start.go:121] virtualization: kvm host
I0615 14:45:09.496967 35758 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x)
I0615 14:45:09.497089 35758 notify.go:169] Checking for updates...
I0615 14:45:09.497486 35758 driver.go:335] Setting default libvirt URI to qemu:///system
I0615 14:45:09.498712 35758 out.go:170] ✨ Using the kvm2 driver based on user configuration
I0615 14:45:09.498722 35758 start.go:279] selected driver: kvm2
I0615 14:45:09.498725 35758 start.go:752] validating driver "kvm2" against <nil>
I0615 14:45:09.498731 35758 start.go:763] status for kvm2: {Installed:true Healthy:false Running:true NeedsImprovement:false Error:libvirt group membership check failed:
error getting current user's GIDs: user: GroupIds requires cgo Reason:PR_KVM_USER_PERMISSION Fix:Check that libvirtd is properly installed and that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) Doc:https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/}
I0615 14:45:09.499768 35758 out.go:170]
W0615 14:45:09.499816 35758 out.go:235] 🚫 Exiting due to PR_KVM_USER_PERMISSION: libvirt group membership check failed:
error getting current user's GIDs: user: GroupIds requires cgo
W0615 14:45:09.499939 35758 out.go:235] 💡 Suggestion: Ensure that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!)
W0615 14:45:09.499974 35758 out.go:235] 📘 Documentation: https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/
W0615 14:45:09.499985 35758 out.go:235] 🍿 Related issues:
W0615 14:45:09.500018 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/5617
W0615 14:45:09.500049 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/10070
🤷 Profile "minikube" not found. Run "minikube profile list" to view all profiles.
👉 To start a cluster, run: "minikube start"
</details>
| non_test | minikube start fails with podman or drivers on minikube claims to run on but i have not had success yet i am on fedora here so perhaps another disto would would better but i ve seen ubuntu fail similarly using the driver steps to reproduce the issue minikube config set driver podman minikube delete minikube start or minikube config set driver minikube delete minikube start full output of minikube logs command for podman last start log file created at running on machine binary built with gc for linux log line format mmdd hh mm ss uuuuuu threadid file line msg out go setting outfile to fd out go isatty isterminal true out go setting errfile to fd out go isatty isterminal true root go updating path home fedora minikube bin out go setting json to false start go hostinfo hostname zdalisv dfw ibm com uptime boottime procs os linux platform fedora platformfamily fedora platformversion kernelversion kernelarch virtualizationsystem kvm virtualizationrole host hostid start go virtualization kvm host out go 😄 minikube on fedora notify go checking for updates driver go setting default libvirt uri to qemu system podman go podman version out go ✨ using the podman driver based on user configuration start go selected driver podman start go validating driver podman against start go status for podman installed true healthy true running false needsimprovement false error reason fix doc cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes start flags go no existing cluster config was found will generate one from the flags start flags go using suggested memory alloc based on sys container start flags go wait components to verify map cni go creating cni manager for cni go cni unnecessary in this configuration recommending no cni start flags go config name minikube keepcontext false embedcerts false minikubeiso kicbaseimage gcr io minikube kicbase memory cpus disksize vmdriver driver podman hyperkitvpnkitsock hyperkitvsockports dockerenv containervolumemounts insecureregistry registrymirror hostonlycidr hypervvirtualswitch hypervuseexternalswitch false hypervexternaladapter kvmnetwork default kvmqemuuri qemu system kvmgpu false kvmhidden false kvmnumacount dockeropt disabledrivermounts false nfsshare nfssharesroot nfsshares uuid novtxcheck false dnsproxy false hostdnsresolver true hostonlynictype virtio natnictype virtio sshipaddress sshuser root sshkey sshport kubernetesconfig kubernetesversion clustername minikube namespace default apiservername minikubeca apiservernames apiserverips dnsdomain cluster local containerruntime docker crisocket networkplugin featuregates servicecidr imagerepository loadbalancerstartip loadbalancerendip customingresscert extraoptions shouldloadcachedimages true enabledefaultcni false cni nodeip nodeport nodename nodes addons map customaddonimages map customaddonregistries map verifycomponents map starthosttimeout scheduledstop exposedports listenaddress network multinoderequested false out go 👍 starting control plane node minikube in cluster minikube cache go beginning downloading kic base image for podman with docker out go 🚜 pulling base image preload go checking if preload exists for version and runtime docker cache go downloading gcr io minikube kicbase to local cache image go checking for gcr io minikube kicbase in local cache directory image go writing gcr io minikube kicbase to local cache preload go status code profile go saving config to home fedora minikube profiles minikube config json lock go writefile acquiring home fedora minikube profiles minikube config json name clock delay timeout cancel cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images docker io kubernetesui metrics scraper exists cache go cache image docker io kubernetesui metrics scraper home fedora minikube cache images docker io kubernetesui metrics scraper took cache go save to tar file docker io kubernetesui metrics scraper home fedora minikube cache images docker io kubernetesui metrics scraper succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube apiserver exists cache go cache image gcr io kube apiserver home fedora minikube cache images gcr io kube apiserver took cache go save to tar file gcr io kube apiserver home fedora minikube cache images gcr io kube apiserver succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube controller manager exists cache go cache image gcr io kube controller manager home fedora minikube cache images gcr io kube controller manager took cache go save to tar file gcr io kube controller manager home fedora minikube cache images gcr io kube controller manager succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube scheduler exists cache go cache image gcr io kube scheduler home fedora minikube cache images gcr io kube scheduler took cache go save to tar file gcr io kube scheduler home fedora minikube cache images gcr io kube scheduler succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube proxy exists cache go cache image gcr io kube proxy home fedora minikube cache images gcr io kube proxy took cache go save to tar file gcr io kube proxy home fedora minikube cache images gcr io kube proxy succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io pause exists cache go cache image gcr io pause home fedora minikube cache images gcr io pause took cache go save to tar file gcr io pause home fedora minikube cache images gcr io pause succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io etcd exists cache go cache image gcr io etcd home fedora minikube cache images gcr io etcd took cache go save to tar file gcr io etcd home fedora minikube cache images gcr io etcd succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io coredns exists cache go cache image gcr io coredns home fedora minikube cache images gcr io coredns took cache go save to tar file gcr io coredns home fedora minikube cache images gcr io coredns succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io minikube storage provisioner exists cache go cache image gcr io minikube storage provisioner home fedora minikube cache images gcr io minikube storage provisioner took cache go save to tar file gcr io minikube storage provisioner home fedora minikube cache images gcr io minikube storage provisioner succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images docker io kubernetesui dashboard exists cache go cache image docker io kubernetesui dashboard home fedora minikube cache images docker io kubernetesui dashboard took cache go save to tar file docker io kubernetesui dashboard home fedora minikube cache images docker io kubernetesui dashboard succeeded cache go successfully saved all images to host disk cache go error downloading kic artifacts not yet implemented see issue cache go successfully downloaded all kic artifacts start go acquiring machines lock for minikube name clock delay timeout cancel start go acquired machines lock for minikube in start go provisioning new machine with config name minikube keepcontext false embedcerts false minikubeiso kicbaseimage gcr io minikube kicbase memory cpus disksize vmdriver driver podman hyperkitvpnkitsock hyperkitvsockports dockerenv containervolumemounts insecureregistry registrymirror hostonlycidr hypervvirtualswitch hypervuseexternalswitch false hypervexternaladapter kvmnetwork default kvmqemuuri qemu system kvmgpu false kvmhidden false kvmnumacount dockeropt disabledrivermounts false nfsshare nfssharesroot nfsshares uuid novtxcheck false dnsproxy false hostdnsresolver true hostonlynictype virtio natnictype virtio sshipaddress sshuser root sshkey sshport kubernetesconfig kubernetesversion clustername minikube namespace default apiservername minikubeca apiservernames apiserverips dnsdomain cluster local containerruntime docker crisocket networkplugin featuregates servicecidr imagerepository loadbalancerstartip loadbalancerendip customingresscert extraoptions shouldloadcachedimages true enabledefaultcni false cni nodeip nodeport nodename nodes addons map customaddonimages map customaddonregistries map verifycomponents map starthosttimeout scheduledstop exposedports listenaddress network multinoderequested false name ip port kubernetesversion controlplane true worker true start go createhost starting for driver podman out go 🔥 creating podman container cpus memory start go libmachine api create for minikube driver podman client go localclient create starting main go libmachine reading certificate data from home fedora minikube certs ca pem main go libmachine decoding pem data main go libmachine parsing certificate main go libmachine reading certificate data from home fedora minikube certs cert pem main go libmachine decoding pem data main go libmachine parsing certificate cli runner go run sudo n podman network inspect minikube format range plugins if eq type bridge index index ipam ranges subnet index index ipam ranges gateway end end network create go found existing network name minikube subnet gateway mtu kic go calculated static ip for the minikube container cli runner go run sudo n podman ps a format names cli runner go run sudo n podman volume create minikube label name minikube sigs io minikube label created by minikube sigs io true oci go successfully created a podman volume minikube cli runner go run sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib cli runner go sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib returned with exit code client go localclient create took ssh runner go run sh c df h var awk nr print cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code start go error running df h var newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go error getting percentage of var that is free newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go duration metric createhost completed in start go releasing machines lock for minikube held for start go error starting host creating host create creating setting up container node preparing volume for minikube container sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib exit status stdout stderr trying to pull gcr io minikube kicbase no image found in manifest list for architecture variant os linux error error choosing an image from manifest list docker gcr io minikube kicbase no image found in manifest list for architecture variant os linux cli runner go run sudo n podman container inspect minikube format state status cli runner go sudo n podman container inspect minikube format state status returned with exit code delete go unable to get host status for minikube assuming it has already been deleted state unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube out go 🤦 starthost failed but will try again creating host create creating setting up container node preparing volume for minikube container sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib exit status stdout stderr trying to pull gcr io minikube kicbase no image found in manifest list for architecture variant os linux error error choosing an image from manifest list docker gcr io minikube kicbase no image found in manifest list for architecture variant os linux start go will try again in seconds start go acquiring machines lock for minikube name clock delay timeout cancel start go acquired machines lock for minikube in start go skipping create using existing machine configuration fix go fixhost starting cli runner go run sudo n podman container inspect minikube format state status cli runner go sudo n podman container inspect minikube format state status returned with exit code fix go recreateifneeded on minikube state err unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube fix go machineexists true err unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube fix go unexpected machine state will restart unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube out go 🔄 restarting existing podman container for minikube cli runner go run sudo n podman start cgroup manager cgroupfs minikube cli runner go sudo n podman start cgroup manager cgroupfs minikube returned with exit code cli runner go run sudo n podman inspect minikube errors go postmortem inspect sudo n podman inspect minikube stdout name minikube driver local mountpoint var lib containers storage volumes minikube data createdat labels created by minikube sigs io true name minikube sigs io minikube scope local options stdout cli runner go run sudo n podman logs timestamps minikube cli runner go sudo n podman logs timestamps minikube returned with exit code errors go failed to get postmortem logs sudo n podman logs timestamps minikube sudo n podman logs timestamps minikube exit status stdout stderr error no container with name or id minikube found no such container cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes errors go postmortem podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes network create go running to gather additional debugging logs cli runner go run sudo n podman network inspect minikube network create go output of stdout cniversion name minikube plugins bridge cni hairpinmode true ipmasq true ipam ranges gateway subnet routes dst type host local isgateway true type bridge capabilities portmappings true type portmap backend type firewall type tuning capabilities aliases true domainname dns podman type dnsname stdout cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes cli runner go run sudo n podman container inspect f networksettings ipaddress minikube cli runner go sudo n podman container inspect f networksettings ipaddress minikube returned with exit code ssh runner go run sh c df h var awk nr print cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code start go error running df h var newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go error getting percentage of var that is free newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube fix go fixhost completed within start go releasing machines lock for minikube held for out go 😿 failed to start podman container running minikube delete may fix it podman inspect ip minikube sudo n podman container inspect f networksettings ipaddress minikube exit status stdout stderr error error inspecting object no such container minikube out go out go ❌ exiting due to guest provision failed to start host podman inspect ip minikube sudo n podman container inspect f minikube exit status stdout stderr error error inspecting object no such container minikube out go out go ╭────────────────────────────────────────────────────────────────────╮ out go │ │ out go │ 😿 if the above advice does not help please let us know │ out go │ 👉 │ out go │ │ out go │ please attach the following file to the github issue │ out go │ home fedora minikube logs laststart txt │ out go │ │ out go ╰────────────────────────────────────────────────────────────────────╯ out go ❌ exiting due to guest status state unknown state minikube sudo n podman container inspect minikube format exit status stdout stderr error error inspecting object no such container minikube ╭───────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 if the above advice does not help please let us know │ │ 👉 │ │ │ │ please attach the following file to the github issue │ │ tmp minikube logs log │ │ │ ╰───────────────────────────────────────────────────────────────────────────╯ full output of minikube logs command for audit command args profile user version start time end time config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc start help minikube fedora tue jun utc tue jun utc start help minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc last start log file created at running on machine binary built with gc for linux log line format mmdd hh mm ss uuuuuu threadid file line msg out go setting outfile to fd out go isatty isterminal true out go setting errfile to fd out go isatty isterminal true root go updating path home fedora minikube bin out go setting json to false start go hostinfo hostname zdalisv dfw ibm com uptime boottime procs os linux platform fedora platformfamily fedora platformversion kernelversion kernelarch virtualizationsystem kvm virtualizationrole host hostid start go virtualization kvm host out go 😄 minikube on fedora notify go checking for updates driver go setting default libvirt uri to qemu system out go ✨ using the driver based on user configuration start go selected driver start go validating driver against start go status for installed true healthy false running true needsimprovement false error libvirt group membership check failed error getting current user s gids user groupids requires cgo reason pr kvm user permission fix check that libvirtd is properly installed and that you are a member of the appropriate libvirt group remember to relogin for group changes to take effect doc out go out go 🚫 exiting due to pr kvm user permission libvirt group membership check failed error getting current user s gids user groupids requires cgo out go 💡 suggestion ensure that you are a member of the appropriate libvirt group remember to relogin for group changes to take effect out go 📘 documentation out go 🍿 related issues out go ▪ out go ▪ 🤷 profile minikube not found run minikube profile list to view all profiles 👉 to start a cluster run minikube start | 0 |
104,757 | 22,749,636,943 | IssuesEvent | 2022-07-07 12:08:25 | Onelinerhub/onelinerhub | https://api.github.com/repos/Onelinerhub/onelinerhub | opened | Short solution needed: "golang regexp" (golang) | help wanted good first issue code golang | Please help us write most modern and shortest code solution for this issue:
**golang regexp** (technology: [golang](https://onelinerhub.com/golang))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution.
3. Link to this issue in comments of pull request. | 1.0 | Short solution needed: "golang regexp" (golang) - Please help us write most modern and shortest code solution for this issue:
**golang regexp** (technology: [golang](https://onelinerhub.com/golang))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution.
3. Link to this issue in comments of pull request. | non_test | short solution needed golang regexp golang please help us write most modern and shortest code solution for this issue golang regexp technology fast way just write the code solution in the comments prefered way create with a new code file inside don t forget to explain solution link to this issue in comments of pull request | 0 |
160,887 | 12,520,964,395 | IssuesEvent | 2020-06-03 16:42:45 | aliasrobotics/RVD | https://api.github.com/repos/aliasrobotics/RVD | opened | Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739 | bandit bug components software robot component: ROS static analysis testing triage version: melodic | ```yaml
{
"id": 1,
"title": "Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:42)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:42)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | 1.0 | Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739 - ```yaml
{
"id": 1,
"title": "Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:42)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:42)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | test | using xml src ros comm roslaunch src roslaunch xmlloader py yaml id title using xml src ros comm roslaunch src roslaunch xmlloader py type bug description high confidence of medium severity bug using xml dom minidom parse to parse untrusted xml data is known to be vulnerable to xml attacks replace xml dom minidom parse with its defusedxml equivalent function or make sure defusedxml defuse stdlib is called src ros comm roslaunch src roslaunch xmlloader py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version melodic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src ros comm roslaunch src roslaunch xmlloader py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation | 1 |
201,101 | 15,173,861,749 | IssuesEvent | 2021-02-13 15:59:53 | XProger/OpenLara | https://api.github.com/repos/XProger/OpenLara | closed | vag decoder crash at titlescreen (OpenBSD) | crash testing | ```
Program received signal SIGABRT, Aborted.
[Switching to thread 414848]
thrkill () at /tmp/-:3
3 /tmp/-: No such file or directory.
in /tmp/-
Current language: auto; currently asm
(gdb) bt
#0 thrkill () at /tmp/-:3
#1 0x000003af38f85b7e in _libc_abort () at /usr/src/lib/libc/stdlib/abort.c:51
#2 0x000003af38f70945 in memcpy (dst0=Variable "dst0" is not available.) at /usr/src/lib/libc/string/memcpy.c:74
#3 0x000003acb6fbfb56 in Sound::VAG::decode (this=0x3af4756d000, frames=0x3af054f4000, count=1024) at sound.h:530
#4 0x000003acb6fc163e in Sound::Sample::render (this=0x3aeee409dc0, frames=0x3af054f4000, count=1024) at sound.h:954
#5 0x000003acb6fc1d46 in Sound::renderChannels (result=0x3af7ab1f000, count=1024, music=false) at sound.h:1064
#6 0x000003acb6fc2244 in Sound::fill (frames=0x3af56c04000, count=1024) at sound.h:1113
#7 0x000003acb7052b21 in sndFill (udata=0x0, stream=0x3aef01e1000 "\\", len=4096) at main.cpp:32
#8 0x000003aedaca7fe3 in ?? () from /usr/local/lib/libSDL2.so.0.8
#9 0x000003aedad11bf0 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8
#10 0x000003aedad95cf7 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8
#11 0x000003af1d7a3e21 in _rthread_start (v=Variable "v" is not available.) at /usr/src/lib/librthread/rthread.c:96
#12 0x000003af38fa8f78 in __tfork_thread () at /usr/src/lib/libc/arch/amd64/sys/tfork_thread.S:77
#13 0x0000000000000000 in ?? ()
(gdb) c
Continuing.
Program terminated with signal SIGABRT, Aborted.
The program no longer exists.
(gdb) q
```
directorytree and more at [pastebin](https://pastebin.com/s0dAqFwP)
Could the `tmp/-: No such file or directory.` be the culprit? how to fix?
| 1.0 | vag decoder crash at titlescreen (OpenBSD) - ```
Program received signal SIGABRT, Aborted.
[Switching to thread 414848]
thrkill () at /tmp/-:3
3 /tmp/-: No such file or directory.
in /tmp/-
Current language: auto; currently asm
(gdb) bt
#0 thrkill () at /tmp/-:3
#1 0x000003af38f85b7e in _libc_abort () at /usr/src/lib/libc/stdlib/abort.c:51
#2 0x000003af38f70945 in memcpy (dst0=Variable "dst0" is not available.) at /usr/src/lib/libc/string/memcpy.c:74
#3 0x000003acb6fbfb56 in Sound::VAG::decode (this=0x3af4756d000, frames=0x3af054f4000, count=1024) at sound.h:530
#4 0x000003acb6fc163e in Sound::Sample::render (this=0x3aeee409dc0, frames=0x3af054f4000, count=1024) at sound.h:954
#5 0x000003acb6fc1d46 in Sound::renderChannels (result=0x3af7ab1f000, count=1024, music=false) at sound.h:1064
#6 0x000003acb6fc2244 in Sound::fill (frames=0x3af56c04000, count=1024) at sound.h:1113
#7 0x000003acb7052b21 in sndFill (udata=0x0, stream=0x3aef01e1000 "\\", len=4096) at main.cpp:32
#8 0x000003aedaca7fe3 in ?? () from /usr/local/lib/libSDL2.so.0.8
#9 0x000003aedad11bf0 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8
#10 0x000003aedad95cf7 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8
#11 0x000003af1d7a3e21 in _rthread_start (v=Variable "v" is not available.) at /usr/src/lib/librthread/rthread.c:96
#12 0x000003af38fa8f78 in __tfork_thread () at /usr/src/lib/libc/arch/amd64/sys/tfork_thread.S:77
#13 0x0000000000000000 in ?? ()
(gdb) c
Continuing.
Program terminated with signal SIGABRT, Aborted.
The program no longer exists.
(gdb) q
```
directorytree and more at [pastebin](https://pastebin.com/s0dAqFwP)
Could the `tmp/-: No such file or directory.` be the culprit? how to fix?
| test | vag decoder crash at titlescreen openbsd program received signal sigabrt aborted thrkill at tmp tmp no such file or directory in tmp current language auto currently asm gdb bt thrkill at tmp in libc abort at usr src lib libc stdlib abort c in memcpy variable is not available at usr src lib libc string memcpy c in sound vag decode this frames count at sound h in sound sample render this frames count at sound h in sound renderchannels result count music false at sound h in sound fill frames count at sound h in sndfill udata stream len at main cpp in from usr local lib so in sdl dynapi entry from usr local lib so in sdl dynapi entry from usr local lib so in rthread start v variable v is not available at usr src lib librthread rthread c in tfork thread at usr src lib libc arch sys tfork thread s in gdb c continuing program terminated with signal sigabrt aborted the program no longer exists gdb q directorytree and more at could the tmp no such file or directory be the culprit how to fix | 1 |
31,532 | 8,708,178,652 | IssuesEvent | 2018-12-06 10:09:36 | indigo-dc/oidc-agent | https://api.github.com/repos/indigo-dc/oidc-agent | closed | Provide possibility to use base64 encoding instead of hex | Build Dependencies LOW Priority enhancement | base64 encoding is supported in lbisodium >= 1.0.14
So we cannot replace hex encoding with base64, but if it is supported by a system we might want to use it | 1.0 | Provide possibility to use base64 encoding instead of hex - base64 encoding is supported in lbisodium >= 1.0.14
So we cannot replace hex encoding with base64, but if it is supported by a system we might want to use it | non_test | provide possibility to use encoding instead of hex encoding is supported in lbisodium so we cannot replace hex encoding with but if it is supported by a system we might want to use it | 0 |
803,855 | 29,191,873,213 | IssuesEvent | 2023-05-19 20:55:06 | fabianabarca/ferias | https://api.github.com/repos/fabianabarca/ferias | closed | Crear nueva tabla para centros de origen | low_priority | Aunque actualmente los centros de origen (según Vavilov) solamente están como opciones en un campo, es posible que en el futuro queramos ampliar la información de cada uno, con datos como una reseña histórica y la ubicación geoespacial (un punto o un polígono) para un mapa, o una imagen. De forma que sería algo como:
```python
class CenterOrigin(models.MODEL):
name =
description =
region = models.PolygonField()
def __str__(self):
return self.name
```
También es posible que sea necesario utilizar la clasificación [actualizada](https://en.wikipedia.org/wiki/Vavilov_center). | 1.0 | Crear nueva tabla para centros de origen - Aunque actualmente los centros de origen (según Vavilov) solamente están como opciones en un campo, es posible que en el futuro queramos ampliar la información de cada uno, con datos como una reseña histórica y la ubicación geoespacial (un punto o un polígono) para un mapa, o una imagen. De forma que sería algo como:
```python
class CenterOrigin(models.MODEL):
name =
description =
region = models.PolygonField()
def __str__(self):
return self.name
```
También es posible que sea necesario utilizar la clasificación [actualizada](https://en.wikipedia.org/wiki/Vavilov_center). | non_test | crear nueva tabla para centros de origen aunque actualmente los centros de origen según vavilov solamente están como opciones en un campo es posible que en el futuro queramos ampliar la información de cada uno con datos como una reseña histórica y la ubicación geoespacial un punto o un polígono para un mapa o una imagen de forma que sería algo como python class centerorigin models model name description region models polygonfield def str self return self name también es posible que sea necesario utilizar la clasificación | 0 |
40,714 | 6,846,066,341 | IssuesEvent | 2017-11-13 10:37:37 | MarinaGarme/GESPRO_PracticaGestionTareas_1718 | https://api.github.com/repos/MarinaGarme/GESPRO_PracticaGestionTareas_1718 | closed | Documentación de crear estructura | documentation | Definir en la wiki una página con las diferentes secciones que deben contener la documentación. | 1.0 | Documentación de crear estructura - Definir en la wiki una página con las diferentes secciones que deben contener la documentación. | non_test | documentación de crear estructura definir en la wiki una página con las diferentes secciones que deben contener la documentación | 0 |
126,676 | 17,093,492,211 | IssuesEvent | 2021-07-08 21:00:51 | PowerShell/PowerShell | https://api.github.com/repos/PowerShell/PowerShell | closed | Allow .ps1xml files in XML 1.1 | Issue-Enhancement Resolution-By Design WG-Cmdlets-Core | To enable colored output in output formats, I can write a Format.ps1xml file that uses ANSI color codes like so:
```xml
<?xml version="1.0" encoding="utf-8"?>
<Configuration>
<ViewDefinitions>
<!-- Resembles the git log output with format=medium (default) -->
<View>
<Name>Medium</Name>
<ViewSelectedBy>
<TypeName>PowerGit.CommitInfo</TypeName>
</ViewSelectedBy>
<ListControl>
<ListEntries>
<ListEntry>
<ListItems>
<ListItem>
<!-- Yellow -->
<Label>[33mSha</Label>
<ScriptBlock>"$($_.Sha)$([char]0x001b)[0m"</ScriptBlock>
</ListItem>
<ListItem>
<Label>Author</Label>
<PropertyName>Author</PropertyName>
</ListItem>
<ListItem>
<Label>Date</Label>
<ScriptBlock>"$($_.Author.When)"</ScriptBlock>
</ListItem>
<ListItem>
<Label>Message</Label>
<PropertyName>Message</PropertyName>
</ListItem>
</ListItems>
</ListEntry>
</ListEntries>
</ListControl>
</View>
</ViewDefinitions>
</Configuration>
```
Note the `[33m`, which is `ESC[33m`, which is the ANSI code for yellow, and `$([char]0x001b)[0m`, which is `ESC[0m`, which is the ANSI reset code.
Together, this will color the whole line yellow:
<img width="452" alt="image" src="https://user-images.githubusercontent.com/10532611/44112434-6167e756-a005-11e8-8c14-895d27beba56.png">
just like in `git log` output:
<img width="345" alt="image" src="https://user-images.githubusercontent.com/10532611/44112448-6b503818-a005-11e8-88da-5e6dabca365d.png">
HOWEVER this is actually not valid XML. Every good XML parser will rightfully complain:
<img width="378" alt="image" src="https://user-images.githubusercontent.com/10532611/44112545-9eedc992-a005-11e8-8aab-0b435ab94fa1.png">
See https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.0
Which means these format files cannot be statically checked, generated, codemodded, etc.
It _is_ valid in XML 1.1: https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.1
But if I change the parsing instruction to
```xml
<?xml version="1.1" encoding="utf-8"?>
```
PowerShell errors:
```powershell
Import-Module : Errors occurred while loading the format data file:
/Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml, Error in file /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml: Version number '1.1' is invalid. Line 1, position 16.
At /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Import-PowerGit.ps1:32 char:1
+ Import-Module -Name (Join-Path -Path $PSScriptRoot -ChildPath 'PowerG ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : InvalidOperation: (:) [Import-Module], RuntimeException
+ FullyQualifiedErrorId : FormatXmlUpdateException,Microsoft.PowerShell.Commands.ImportModuleCommand
```
It effectively already parses XML 1.1 fine though, it just complains about the version.
| 1.0 | Allow .ps1xml files in XML 1.1 - To enable colored output in output formats, I can write a Format.ps1xml file that uses ANSI color codes like so:
```xml
<?xml version="1.0" encoding="utf-8"?>
<Configuration>
<ViewDefinitions>
<!-- Resembles the git log output with format=medium (default) -->
<View>
<Name>Medium</Name>
<ViewSelectedBy>
<TypeName>PowerGit.CommitInfo</TypeName>
</ViewSelectedBy>
<ListControl>
<ListEntries>
<ListEntry>
<ListItems>
<ListItem>
<!-- Yellow -->
<Label>[33mSha</Label>
<ScriptBlock>"$($_.Sha)$([char]0x001b)[0m"</ScriptBlock>
</ListItem>
<ListItem>
<Label>Author</Label>
<PropertyName>Author</PropertyName>
</ListItem>
<ListItem>
<Label>Date</Label>
<ScriptBlock>"$($_.Author.When)"</ScriptBlock>
</ListItem>
<ListItem>
<Label>Message</Label>
<PropertyName>Message</PropertyName>
</ListItem>
</ListItems>
</ListEntry>
</ListEntries>
</ListControl>
</View>
</ViewDefinitions>
</Configuration>
```
Note the `[33m`, which is `ESC[33m`, which is the ANSI code for yellow, and `$([char]0x001b)[0m`, which is `ESC[0m`, which is the ANSI reset code.
Together, this will color the whole line yellow:
<img width="452" alt="image" src="https://user-images.githubusercontent.com/10532611/44112434-6167e756-a005-11e8-8c14-895d27beba56.png">
just like in `git log` output:
<img width="345" alt="image" src="https://user-images.githubusercontent.com/10532611/44112448-6b503818-a005-11e8-88da-5e6dabca365d.png">
HOWEVER this is actually not valid XML. Every good XML parser will rightfully complain:
<img width="378" alt="image" src="https://user-images.githubusercontent.com/10532611/44112545-9eedc992-a005-11e8-8aab-0b435ab94fa1.png">
See https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.0
Which means these format files cannot be statically checked, generated, codemodded, etc.
It _is_ valid in XML 1.1: https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.1
But if I change the parsing instruction to
```xml
<?xml version="1.1" encoding="utf-8"?>
```
PowerShell errors:
```powershell
Import-Module : Errors occurred while loading the format data file:
/Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml, Error in file /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml: Version number '1.1' is invalid. Line 1, position 16.
At /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Import-PowerGit.ps1:32 char:1
+ Import-Module -Name (Join-Path -Path $PSScriptRoot -ChildPath 'PowerG ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : InvalidOperation: (:) [Import-Module], RuntimeException
+ FullyQualifiedErrorId : FormatXmlUpdateException,Microsoft.PowerShell.Commands.ImportModuleCommand
```
It effectively already parses XML 1.1 fine though, it just complains about the version.
| non_test | allow files in xml to enable colored output in output formats i can write a format file that uses ansi color codes like so xml medium powergit commitinfo sha author author date author when message message note the which is esc which is the ansi reset code together this will color the whole line yellow img width alt image src just like in git log output img width alt image src however this is actually not valid xml every good xml parser will rightfully complain img width alt image src see which means these format files cannot be statically checked generated codemodded etc it is valid in xml but if i change the parsing instruction to xml powershell errors powershell import module errors occurred while loading the format data file users felix src github com felixfbecker powergit powergit formats powergit commitinfo formats error in file users felix src github com felixfbecker powergit powergit formats powergit commitinfo formats version number is invalid line position at users felix src github com felixfbecker powergit powergit import powergit char import module name join path path psscriptroot childpath powerg categoryinfo invalidoperation runtimeexception fullyqualifiederrorid formatxmlupdateexception microsoft powershell commands importmodulecommand it effectively already parses xml fine though it just complains about the version | 0 |
110,607 | 9,462,290,017 | IssuesEvent | 2019-04-17 15:08:45 | LIBCAS/INDIHU-Exhibition | https://api.github.com/repos/LIBCAS/INDIHU-Exhibition | closed | Náhled výstavy | waiting for test | Opět nám nefunguje (na několika počítačích i v různých prohlížečích) náhled výstavy. Snad podobný problém již řešen #45 a #5 . Jedná se o výstavu: http://inqooltest.libj.cas.cz/expo/ac131145-410d-4519-a36b-1ffac4925b3c/structure
| 1.0 | Náhled výstavy - Opět nám nefunguje (na několika počítačích i v různých prohlížečích) náhled výstavy. Snad podobný problém již řešen #45 a #5 . Jedná se o výstavu: http://inqooltest.libj.cas.cz/expo/ac131145-410d-4519-a36b-1ffac4925b3c/structure
| test | náhled výstavy opět nám nefunguje na několika počítačích i v různých prohlížečích náhled výstavy snad podobný problém již řešen a jedná se o výstavu | 1 |
148,576 | 11,858,139,075 | IssuesEvent | 2020-03-25 10:52:05 | clarity-h2020/csis-technical-validation | https://api.github.com/repos/clarity-h2020/csis-technical-validation | closed | Setup cypress.io UI integration Tests | testing | Setup cypress.io UI Tests for CSIS
- [x] ~~[Fix Jenkins issues](https://github.com/clarity-h2020/csis/issues/8#issuecomment-575546178)~~
- [x] ~~create new GitHub Jenkins users~~
- [x] create 1st simple UI tests (Jenkins Job)
| 1.0 | Setup cypress.io UI integration Tests - Setup cypress.io UI Tests for CSIS
- [x] ~~[Fix Jenkins issues](https://github.com/clarity-h2020/csis/issues/8#issuecomment-575546178)~~
- [x] ~~create new GitHub Jenkins users~~
- [x] create 1st simple UI tests (Jenkins Job)
| test | setup cypress io ui integration tests setup cypress io ui tests for csis create new github jenkins users create simple ui tests jenkins job | 1 |
105,027 | 16,623,626,777 | IssuesEvent | 2021-06-03 06:45:37 | Thanraj/OpenSSL_1.0.1 | https://api.github.com/repos/Thanraj/OpenSSL_1.0.1 | opened | CVE-2014-3569 (Medium) detected in opensslOpenSSL_1_0_1 | security vulnerability | ## CVE-2014-3569 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ssl23_get_client_hello function in s23_srvr.c in OpenSSL 0.9.8zc, 1.0.0o, and 1.0.1j does not properly handle attempts to use unsupported protocols, which allows remote attackers to cause a denial of service (NULL pointer dereference and daemon crash) via an unexpected handshake, as demonstrated by an SSLv3 handshake to a no-ssl3 application with certain error handling. NOTE: this issue became relevant after the CVE-2014-3568 fix.
<p>Publish Date: 2014-12-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3569>CVE-2014-3569</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569</a></p>
<p>Release Date: 2014-12-24</p>
<p>Fix Resolution: OpenSSL_1_0_2a</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2014-3569 (Medium) detected in opensslOpenSSL_1_0_1 - ## CVE-2014-3569 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ssl23_get_client_hello function in s23_srvr.c in OpenSSL 0.9.8zc, 1.0.0o, and 1.0.1j does not properly handle attempts to use unsupported protocols, which allows remote attackers to cause a denial of service (NULL pointer dereference and daemon crash) via an unexpected handshake, as demonstrated by an SSLv3 handshake to a no-ssl3 application with certain error handling. NOTE: this issue became relevant after the CVE-2014-3568 fix.
<p>Publish Date: 2014-12-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3569>CVE-2014-3569</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569</a></p>
<p>Release Date: 2014-12-24</p>
<p>Fix Resolution: OpenSSL_1_0_2a</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in opensslopenssl cve medium severity vulnerability vulnerable library opensslopenssl akamai fork of openssl master library home page a href found in head commit a href found in base branch master vulnerable source files openssl ssl srvr c openssl ssl srvr c openssl ssl srvr c vulnerability details the get client hello function in srvr c in openssl and does not properly handle attempts to use unsupported protocols which allows remote attackers to cause a denial of service null pointer dereference and daemon crash via an unexpected handshake as demonstrated by an handshake to a no application with certain error handling note this issue became relevant after the cve fix publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution openssl step up your open source security game with whitesource | 0 |
81,523 | 10,240,698,391 | IssuesEvent | 2019-08-19 21:31:15 | ProjectEvergreen/greenwood | https://api.github.com/repos/ProjectEvergreen/greenwood | opened | 404 page not building and serializing correctly | bug documentation question | ## Type of Change
- [ ] New Feature Request
- [ ] Documentation / Website
- [ ] Improvement / Suggestion
- [x] Bug
- [ ] Other (please clarify below)
## Summary
Observed a couple issues with 404 page, the default anyway. Observed from [my own project](https://github.com/thegreenhouseio/www.thegreenhouse.io/issues/112) but also in getting started companion repo. Probably exists in _www/_ too.
**develop**
1. `<title>` is default value, _My App_. Not sure if there could be more missing config items, like `meta` or others?
<img width="310" alt="Screen Shot 2019-08-19 at 5 29 38 PM" src="https://user-images.githubusercontent.com/895923/63300781-09211200-c2a7-11e9-8efb-bae93e65c0af.png">
**build**
1. It doesn't seem to exist at all?
<img width="703" alt="Screen Shot 2019-08-19 at 5 29 33 PM" src="https://user-images.githubusercontent.com/895923/63300768-032b3100-c2a7-11e9-90d5-c780fe0826d7.png">
## Details
Maybe this is more of a need around documentation as well? Not 100% clear what the expectations are. Do users make their own 404.md? | 1.0 | 404 page not building and serializing correctly - ## Type of Change
- [ ] New Feature Request
- [ ] Documentation / Website
- [ ] Improvement / Suggestion
- [x] Bug
- [ ] Other (please clarify below)
## Summary
Observed a couple issues with 404 page, the default anyway. Observed from [my own project](https://github.com/thegreenhouseio/www.thegreenhouse.io/issues/112) but also in getting started companion repo. Probably exists in _www/_ too.
**develop**
1. `<title>` is default value, _My App_. Not sure if there could be more missing config items, like `meta` or others?
<img width="310" alt="Screen Shot 2019-08-19 at 5 29 38 PM" src="https://user-images.githubusercontent.com/895923/63300781-09211200-c2a7-11e9-8efb-bae93e65c0af.png">
**build**
1. It doesn't seem to exist at all?
<img width="703" alt="Screen Shot 2019-08-19 at 5 29 33 PM" src="https://user-images.githubusercontent.com/895923/63300768-032b3100-c2a7-11e9-90d5-c780fe0826d7.png">
## Details
Maybe this is more of a need around documentation as well? Not 100% clear what the expectations are. Do users make their own 404.md? | non_test | page not building and serializing correctly type of change new feature request documentation website improvement suggestion bug other please clarify below summary observed a couple issues with page the default anyway observed from but also in getting started companion repo probably exists in www too develop is default value my app not sure if there could be more missing config items like meta or others img width alt screen shot at pm src build it doesn t seem to exist at all img width alt screen shot at pm src details maybe this is more of a need around documentation as well not clear what the expectations are do users make their own md | 0 |
102,983 | 8,872,905,759 | IssuesEvent | 2019-01-11 16:35:17 | LiskHQ/lisk-hub | https://api.github.com/repos/LiskHQ/lisk-hub | closed | Register integration tests fail randomly | type: test | Sometimes this test fail:
```
FAILED TESTS:
@integration: Register
Scenario: should allow to create a second passphrase
✖ Then I should be logged in
HeadlessChrome 70.0.3538 (Linux 0.0.0)
AssertionError: expected activePeerSet to have been called with arguments match(passphrase: punch indicate carbon slush gospel ask museum file actress spawn merit mechanic)
at Helper.checkIfRegistrationConfirmed (webpack:///test/integration/register.test.js:76:44 <- src/tests.js:403066:58)
at Context.<anonymous> (webpack:///test/integration/register.test.js:133:52 <- src/tests.js:403154:21)
at Context.sync (webpack:///node_modules/mocha-steps/lib/step.js:29:0 <- src/tests.js:89665:24)
```
https://jenkins.lisk.io/job/lisk-hub/job/1.5.0/23/consoleFull | 1.0 | Register integration tests fail randomly - Sometimes this test fail:
```
FAILED TESTS:
@integration: Register
Scenario: should allow to create a second passphrase
✖ Then I should be logged in
HeadlessChrome 70.0.3538 (Linux 0.0.0)
AssertionError: expected activePeerSet to have been called with arguments match(passphrase: punch indicate carbon slush gospel ask museum file actress spawn merit mechanic)
at Helper.checkIfRegistrationConfirmed (webpack:///test/integration/register.test.js:76:44 <- src/tests.js:403066:58)
at Context.<anonymous> (webpack:///test/integration/register.test.js:133:52 <- src/tests.js:403154:21)
at Context.sync (webpack:///node_modules/mocha-steps/lib/step.js:29:0 <- src/tests.js:89665:24)
```
https://jenkins.lisk.io/job/lisk-hub/job/1.5.0/23/consoleFull | test | register integration tests fail randomly sometimes this test fail failed tests integration register scenario should allow to create a second passphrase ✖ then i should be logged in headlesschrome linux assertionerror expected activepeerset to have been called with arguments match passphrase punch indicate carbon slush gospel ask museum file actress spawn merit mechanic at helper checkifregistrationconfirmed webpack test integration register test js src tests js at context webpack test integration register test js src tests js at context sync webpack node modules mocha steps lib step js src tests js | 1 |
33,604 | 27,616,422,801 | IssuesEvent | 2023-03-09 19:44:59 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | closed | Upgrade ESLint to latest | P1 Type: Infrastructure Exp: SP | ## Feature Description
We currently use ESLint v7 but not the latest version. Because of this, an older version of lodash is installed. See https://github.com/google/site-kit-wp/security/dependabot/34
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* ESLint should be upgraded to the latest 7.x version (currently 7.32.0)
## Implementation Brief
* Upgrade `eslint` to `^7.32.0`
### Test Coverage
* No changes
## QA Brief
* This only affects development tooling, so no additional QA should be necessary. We mainly need to ensure ESLint checks run as expected after upgrading, which should be observable in CR
## Changelog entry
* N/A
| 1.0 | Upgrade ESLint to latest - ## Feature Description
We currently use ESLint v7 but not the latest version. Because of this, an older version of lodash is installed. See https://github.com/google/site-kit-wp/security/dependabot/34
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* ESLint should be upgraded to the latest 7.x version (currently 7.32.0)
## Implementation Brief
* Upgrade `eslint` to `^7.32.0`
### Test Coverage
* No changes
## QA Brief
* This only affects development tooling, so no additional QA should be necessary. We mainly need to ensure ESLint checks run as expected after upgrading, which should be observable in CR
## Changelog entry
* N/A
| non_test | upgrade eslint to latest feature description we currently use eslint but not the latest version because of this an older version of lodash is installed see do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria eslint should be upgraded to the latest x version currently implementation brief upgrade eslint to test coverage no changes qa brief this only affects development tooling so no additional qa should be necessary we mainly need to ensure eslint checks run as expected after upgrading which should be observable in cr changelog entry n a | 0 |
286,644 | 24,767,557,190 | IssuesEvent | 2022-10-22 18:25:33 | wasabee-project/Wasabee-IITC | https://api.github.com/repos/wasabee-project/Wasabee-IITC | closed | Filter tasks sharing some properties and bulk setter | In Testing | Some feature requests have in common the need to color all links sharing a property, like the source portal or the assigned agent.
I have some ideas for this:
1. the #318 branch to select a subset of the links in the checklist and color them
2. integrate buttons (for instance) in the link list (wrt a portal) to color all inbound or outbound links
3. create a UI to give (client-side) a color to each agent and use this data to color implicitly the rendering
4. rule base coloring (rendering won't use the `color` attribute for a link) (see https://github.com/wasabee-project/Wasabee-IITC/commit/9d9d1b672afb539e508454835545a6df30958c21 for pin color)
**Properties:**
- by assignment
- by target
- by source
- by zone
Related: #318 #335 #350 #191
| 1.0 | Filter tasks sharing some properties and bulk setter - Some feature requests have in common the need to color all links sharing a property, like the source portal or the assigned agent.
I have some ideas for this:
1. the #318 branch to select a subset of the links in the checklist and color them
2. integrate buttons (for instance) in the link list (wrt a portal) to color all inbound or outbound links
3. create a UI to give (client-side) a color to each agent and use this data to color implicitly the rendering
4. rule base coloring (rendering won't use the `color` attribute for a link) (see https://github.com/wasabee-project/Wasabee-IITC/commit/9d9d1b672afb539e508454835545a6df30958c21 for pin color)
**Properties:**
- by assignment
- by target
- by source
- by zone
Related: #318 #335 #350 #191
| test | filter tasks sharing some properties and bulk setter some feature requests have in common the need to color all links sharing a property like the source portal or the assigned agent i have some ideas for this the branch to select a subset of the links in the checklist and color them integrate buttons for instance in the link list wrt a portal to color all inbound or outbound links create a ui to give client side a color to each agent and use this data to color implicitly the rendering rule base coloring rendering won t use the color attribute for a link see for pin color properties by assignment by target by source by zone related | 1 |
286,395 | 8,787,624,891 | IssuesEvent | 2018-12-20 19:18:00 | containous/traefik | https://api.github.com/repos/containous/traefik | closed | Traefik killed by OOM killer in Kubernetes | area/provider/k8s kind/bug/confirmed priority/P2 | ### Do you want to request a *feature* or report a *bug*?
Bug
### What did you do?
- Running Traefik as Deployment (ingress controller) in K8s
- Having bunch of Services targeting no Pods (thus having no related Endpoints object)
- Our setup is specific in way that we downscale tens of deployments to 0 on some schedule to not waste resources of K8s cluster.
### What did you expect to see?
- Traefik to work and reply with error when tried to access such a service
### What did you see instead?
- Traefik leaking memory and eventually being killed by OOM killer
### Output of `traefik version`: (_What version of Traefik are you using?_)
1.7.4
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
defaultEntryPoints = ["http","https"]
[entryPoints]
[entryPoints.http]
address = ":80"
compress = true
[entryPoints.http.proxyProtocol]
trustedIPs = ["xxxx"]
[entryPoints.https]
address = ":443"
compress = true
[entryPoints.https.proxyProtocol]
trustedIPs = ["xxxx"]
[entryPoints.https.tls]
minVersion = "VersionTLS11"
cipherSuites = [
"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305",
"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA",
"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA",
"TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA",
"TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA"
]
[respondingTimeouts]
idleTimeout = "900s"
[lifeCycle]
requestAcceptGraceTimeout = "10s"
graceTimeOut = "10s"
[api]
[rest]
[kubernetes]
[kubernetes.ingressEndpoint]
hostname = "xxxx.xxxx"
[metrics]
[metrics.prometheus]
[accessLog]
```
### Logs
```
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
```
Tons of those ^
I suspect it being a memory leak in watcher watching the Services/Endpoints as this memory growth is specific just in this case, on our other clusters where we don't downscale Deployments like this Traefik works just fine. | 1.0 | Traefik killed by OOM killer in Kubernetes - ### Do you want to request a *feature* or report a *bug*?
Bug
### What did you do?
- Running Traefik as Deployment (ingress controller) in K8s
- Having bunch of Services targeting no Pods (thus having no related Endpoints object)
- Our setup is specific in way that we downscale tens of deployments to 0 on some schedule to not waste resources of K8s cluster.
### What did you expect to see?
- Traefik to work and reply with error when tried to access such a service
### What did you see instead?
- Traefik leaking memory and eventually being killed by OOM killer
### Output of `traefik version`: (_What version of Traefik are you using?_)
1.7.4
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
defaultEntryPoints = ["http","https"]
[entryPoints]
[entryPoints.http]
address = ":80"
compress = true
[entryPoints.http.proxyProtocol]
trustedIPs = ["xxxx"]
[entryPoints.https]
address = ":443"
compress = true
[entryPoints.https.proxyProtocol]
trustedIPs = ["xxxx"]
[entryPoints.https.tls]
minVersion = "VersionTLS11"
cipherSuites = [
"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305",
"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA",
"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA",
"TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA",
"TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA"
]
[respondingTimeouts]
idleTimeout = "900s"
[lifeCycle]
requestAcceptGraceTimeout = "10s"
graceTimeOut = "10s"
[api]
[rest]
[kubernetes]
[kubernetes.ingressEndpoint]
hostname = "xxxx.xxxx"
[metrics]
[metrics.prometheus]
[accessLog]
```
### Logs
```
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx"
```
Tons of those ^
I suspect it being a memory leak in watcher watching the Services/Endpoints as this memory growth is specific just in this case, on our other clusters where we don't downscale Deployments like this Traefik works just fine. | non_test | traefik killed by oom killer in kubernetes do you want to request a feature or report a bug bug what did you do running traefik as deployment ingress controller in having bunch of services targeting no pods thus having no related endpoints object our setup is specific in way that we downscale tens of deployments to on some schedule to not waste resources of cluster what did you expect to see traefik to work and reply with error when tried to access such a service what did you see instead traefik leaking memory and eventually being killed by oom killer output of traefik version what version of traefik are you using what is your environment configuration arguments toml provider platform toml defaultentrypoints address compress true trustedips address compress true trustedips minversion ciphersuites tls ecdhe rsa with aes gcm tls ecdhe rsa with tls ecdhe rsa with aes cbc sha tls ecdhe rsa with aes cbc sha tls ecdhe ecdsa with aes cbc sha tls ecdhe ecdsa with aes cbc sha idletimeout requestacceptgracetimeout gracetimeout hostname xxxx xxxx logs time level warning msg endpoints not available for xxx time level warning msg endpoints not available for xxx time level warning msg endpoints not available for xxx tons of those i suspect it being a memory leak in watcher watching the services endpoints as this memory growth is specific just in this case on our other clusters where we don t downscale deployments like this traefik works just fine | 0 |
197,196 | 14,911,804,112 | IssuesEvent | 2021-01-22 11:39:11 | DanielMurphy22/SmokeTests | https://api.github.com/repos/DanielMurphy22/SmokeTests | closed | Ubuntu Plotting Smoke Tests | Linux Only Manual Tests |
Before testing:
- Check this testing issue relates to the OS you will test on.
- If unassigned, please assign yourself as for a normal Github issue.
- Please run these tests on the release package of Mantid; **not a locally built version**.
Afterwards:
- Comment below with any issues you came across.
- If no issues were found, or they are now all resolved, please close the testing issue.
- Check the master issue for this OS for other unassigned smoke tests.
If you have any questions please contact the creator of this issue.
:chart_with_upwards_trend: :bar_chart:
* 1D plotting:[instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#displaying-1d-data)
- [ ] Simple plot
- [ ] Another way to plot
- [ ] Adding curves to existing plots
- [ ] Also, test out [waterfall](https://docs.mantidproject.org/nightly/plotting/WaterfallPlotsHelp.html#waterfall-plots) and [tiled]( https://docs.mantidproject.org/nightly/plotting/1DPlotsHelp.html#tiled-plots)
- [ ] Check Toolbar buttons
* 2D plotting: [instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/04_displaying_2D_data.html#displaying-2d-data)
- [ ] Plot all spectra
- [ ] Change colour map
- [ ] [Contour plot](https://docs.mantidproject.org/nightly/plotting/ColorfillPlotsHelp.html#contour-plots) (under 3D menu)
- [ ] Check Toolbar buttons
* [3D plotting](https://docs.mantidproject.org/nightly/plotting/3DPlotsHelp.html):
- Load some data eg `LOQ7041` from the ISIS sample data
- [ ] 3D surface
- [ ] 3D wire frame
- [ ] Check Toolbar buttons
- [ ] ## Sliceviewer (further instructions and data [here](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html)) In particular try editing the data in a workspace while it is open in Sliceviewer! ([step 10](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html#alter-the-underlying-workspace)) | 1.0 | Ubuntu Plotting Smoke Tests -
Before testing:
- Check this testing issue relates to the OS you will test on.
- If unassigned, please assign yourself as for a normal Github issue.
- Please run these tests on the release package of Mantid; **not a locally built version**.
Afterwards:
- Comment below with any issues you came across.
- If no issues were found, or they are now all resolved, please close the testing issue.
- Check the master issue for this OS for other unassigned smoke tests.
If you have any questions please contact the creator of this issue.
:chart_with_upwards_trend: :bar_chart:
* 1D plotting:[instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#displaying-1d-data)
- [ ] Simple plot
- [ ] Another way to plot
- [ ] Adding curves to existing plots
- [ ] Also, test out [waterfall](https://docs.mantidproject.org/nightly/plotting/WaterfallPlotsHelp.html#waterfall-plots) and [tiled]( https://docs.mantidproject.org/nightly/plotting/1DPlotsHelp.html#tiled-plots)
- [ ] Check Toolbar buttons
* 2D plotting: [instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/04_displaying_2D_data.html#displaying-2d-data)
- [ ] Plot all spectra
- [ ] Change colour map
- [ ] [Contour plot](https://docs.mantidproject.org/nightly/plotting/ColorfillPlotsHelp.html#contour-plots) (under 3D menu)
- [ ] Check Toolbar buttons
* [3D plotting](https://docs.mantidproject.org/nightly/plotting/3DPlotsHelp.html):
- Load some data eg `LOQ7041` from the ISIS sample data
- [ ] 3D surface
- [ ] 3D wire frame
- [ ] Check Toolbar buttons
- [ ] ## Sliceviewer (further instructions and data [here](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html)) In particular try editing the data in a workspace while it is open in Sliceviewer! ([step 10](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html#alter-the-underlying-workspace)) | test | ubuntu plotting smoke tests before testing check this testing issue relates to the os you will test on if unassigned please assign yourself as for a normal github issue please run these tests on the release package of mantid not a locally built version afterwards comment below with any issues you came across if no issues were found or they are now all resolved please close the testing issue check the master issue for this os for other unassigned smoke tests if you have any questions please contact the creator of this issue chart with upwards trend bar chart plotting simple plot another way to plot adding curves to existing plots also test out and check toolbar buttons plotting plot all spectra change colour map under menu check toolbar buttons load some data eg from the isis sample data surface wire frame check toolbar buttons sliceviewer further instructions and data in particular try editing the data in a workspace while it is open in sliceviewer | 1 |
243,860 | 20,593,929,781 | IssuesEvent | 2022-03-05 07:11:37 | freelearn2000/Project1 | https://api.github.com/repos/freelearn2000/Project1 | opened | Automated Issue - Node App Tests/Coverage Failed | test | Someone just pushed, oh no! Here's who did it: NadaMusthafa.
[Check test report here.](https://freelearn2000.github.io/Project1/nodeapp/index.html)
[Check coverage report here.](https://freelearn2000.github.io/Project1/nodeapp/coverage/lcov-report/index.html)
[Check lint report here.](https://freelearn2000.github.io/Project1/nodeapp/reports/lint-results.html)
| 1.0 | Automated Issue - Node App Tests/Coverage Failed - Someone just pushed, oh no! Here's who did it: NadaMusthafa.
[Check test report here.](https://freelearn2000.github.io/Project1/nodeapp/index.html)
[Check coverage report here.](https://freelearn2000.github.io/Project1/nodeapp/coverage/lcov-report/index.html)
[Check lint report here.](https://freelearn2000.github.io/Project1/nodeapp/reports/lint-results.html)
| test | automated issue node app tests coverage failed someone just pushed oh no here s who did it nadamusthafa | 1 |
218,462 | 16,761,190,022 | IssuesEvent | 2021-06-13 20:27:55 | AlfredoCarias/Examen-Final-Analisis-de-sistemas | https://api.github.com/repos/AlfredoCarias/Examen-Final-Analisis-de-sistemas | closed | Realizacion del documento inicial del sistema | documentation | Se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo. | 1.0 | Realizacion del documento inicial del sistema - Se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo. | non_test | realizacion del documento inicial del sistema se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo | 0 |
31,410 | 4,705,774,368 | IssuesEvent | 2016-10-13 15:21:11 | PulpQE/pulp-smash | https://api.github.com/repos/PulpQE/pulp-smash | opened | Test full sync with different scenarios- | Issue Type: Discussion Issue Type: Test Case Status: Needs Work | https://pulp.plan.io/issues/1983
The sync should be full in case one of these answers are true:
1. Did the config change in any way since the last sync?
- A new field should be added to the Importer model 'last_updated'. As a consequence a migration should be written.
2. Has content been removed since the last sync?
3. If there was something specified in the override_config( because it changes current sync).
- A new field should be added to the Importer model 'last_override_config'. As a consequence a migration should be written.
4. Force-full option was specified. | 1.0 | Test full sync with different scenarios- - https://pulp.plan.io/issues/1983
The sync should be full in case one of these answers are true:
1. Did the config change in any way since the last sync?
- A new field should be added to the Importer model 'last_updated'. As a consequence a migration should be written.
2. Has content been removed since the last sync?
3. If there was something specified in the override_config( because it changes current sync).
- A new field should be added to the Importer model 'last_override_config'. As a consequence a migration should be written.
4. Force-full option was specified. | test | test full sync with different scenarios the sync should be full in case one of these answers are true did the config change in any way since the last sync a new field should be added to the importer model last updated as a consequence a migration should be written has content been removed since the last sync if there was something specified in the override config because it changes current sync a new field should be added to the importer model last override config as a consequence a migration should be written force full option was specified | 1 |
332,522 | 29,482,290,139 | IssuesEvent | 2023-06-02 06:59:26 | mratsim/constantine | https://api.github.com/repos/mratsim/constantine | opened | Reactivate 32-bit CI | testing :passport_control: | 32-bit CI is failing recently. It may be fixed whern a new Github CI Agent is out with more recent packages that avoid the ``you have held broken packages`` error.

```
Run sudo dpkg --add-architecture i386
sudo dpkg --add-architecture i386
sudo apt-fast update -qq
# Try to fix "E: Unable to correct problems, you have held broken packages."
sudo apt-fast clean
sudo DEBIAN_FRONTEND='noninteractive' apt-fast install \
--no-install-recommends -yq \
gcc-multilib g++-multilib \
libssl-dev:i386 libgmp-dev:i386
mkdir -p external/bin
cat << EOF > external/bin/gcc
#!/bin/bash
exec $(which gcc) -m32 "\$@"
EOF
cat << EOF > external/bin/g++
#!/bin/bash
exec $(which g++) -m32 "\$@"
EOF
chmod 755 external/bin/{gcc,g++}
echo '/home/runner/work/constantine/constantine/external/bin' >> $GITHUB_PATH
shell: /usr/bin/bash -e {0}
[apt-fast 06:24:01]
[apt-fast 06:24:01]Working... this may take a while.
E: Unable to correct problems, you have held broken packages.
[apt-fast 06:24:01]Package manager quit with exit code.
Error: Process completed with exit code 100.
``` | 1.0 | Reactivate 32-bit CI - 32-bit CI is failing recently. It may be fixed whern a new Github CI Agent is out with more recent packages that avoid the ``you have held broken packages`` error.

```
Run sudo dpkg --add-architecture i386
sudo dpkg --add-architecture i386
sudo apt-fast update -qq
# Try to fix "E: Unable to correct problems, you have held broken packages."
sudo apt-fast clean
sudo DEBIAN_FRONTEND='noninteractive' apt-fast install \
--no-install-recommends -yq \
gcc-multilib g++-multilib \
libssl-dev:i386 libgmp-dev:i386
mkdir -p external/bin
cat << EOF > external/bin/gcc
#!/bin/bash
exec $(which gcc) -m32 "\$@"
EOF
cat << EOF > external/bin/g++
#!/bin/bash
exec $(which g++) -m32 "\$@"
EOF
chmod 755 external/bin/{gcc,g++}
echo '/home/runner/work/constantine/constantine/external/bin' >> $GITHUB_PATH
shell: /usr/bin/bash -e {0}
[apt-fast 06:24:01]
[apt-fast 06:24:01]Working... this may take a while.
E: Unable to correct problems, you have held broken packages.
[apt-fast 06:24:01]Package manager quit with exit code.
Error: Process completed with exit code 100.
``` | test | reactivate bit ci bit ci is failing recently it may be fixed whern a new github ci agent is out with more recent packages that avoid the you have held broken packages error run sudo dpkg add architecture sudo dpkg add architecture sudo apt fast update qq try to fix e unable to correct problems you have held broken packages sudo apt fast clean sudo debian frontend noninteractive apt fast install no install recommends yq gcc multilib g multilib libssl dev libgmp dev mkdir p external bin cat external bin gcc bin bash exec which gcc eof cat external bin g bin bash exec which g eof chmod external bin gcc g echo home runner work constantine constantine external bin github path shell usr bin bash e working this may take a while e unable to correct problems you have held broken packages package manager quit with exit code error process completed with exit code | 1 |
213,784 | 16,537,103,671 | IssuesEvent | 2021-05-27 13:03:10 | finos/waltz | https://api.github.com/repos/finos/waltz | closed | Survey Tweaks: minor fixes | QoL fixed (test & close) |
**Instances Summary**
- [x] color code the status
**Instances**
- [x] when viewing make it look nicer
- [x] editing, widgets should line up better
| 1.0 | Survey Tweaks: minor fixes -
**Instances Summary**
- [x] color code the status
**Instances**
- [x] when viewing make it look nicer
- [x] editing, widgets should line up better
| test | survey tweaks minor fixes instances summary color code the status instances when viewing make it look nicer editing widgets should line up better | 1 |
52,958 | 10,964,450,029 | IssuesEvent | 2019-11-27 22:39:09 | The-Squadratics/openMRS_security_project | https://api.github.com/repos/The-Squadratics/openMRS_security_project | closed | Code Analysis for SSE - Task 0 - Create Working Document | Code Review | ### Code Analysis for SSE - Task 0 - Project documentation
The intent of the task is to create the working document and initial tasks for the Code Analysis for SSE assignment.
### Deliverable
* Create a new markdown document in which we'll track the progress of this assignment
* Create the initial task issues for this assignment | 1.0 | Code Analysis for SSE - Task 0 - Create Working Document - ### Code Analysis for SSE - Task 0 - Project documentation
The intent of the task is to create the working document and initial tasks for the Code Analysis for SSE assignment.
### Deliverable
* Create a new markdown document in which we'll track the progress of this assignment
* Create the initial task issues for this assignment | non_test | code analysis for sse task create working document code analysis for sse task project documentation the intent of the task is to create the working document and initial tasks for the code analysis for sse assignment deliverable create a new markdown document in which we ll track the progress of this assignment create the initial task issues for this assignment | 0 |
466,133 | 13,397,419,703 | IssuesEvent | 2020-09-03 11:34:10 | jimbertools/timby | https://api.github.com/repos/jimbertools/timby | closed | Session close message and actual session closing is out of sync. | bug priority_medium | Sometimes when people are on spotty connections, they have issues with timby saying your session is closed, but then they still send a request and update their session before it closes.
These need to be in sync so the "your session is closed" message is only sent when it's actually closed.

| 1.0 | Session close message and actual session closing is out of sync. - Sometimes when people are on spotty connections, they have issues with timby saying your session is closed, but then they still send a request and update their session before it closes.
These need to be in sync so the "your session is closed" message is only sent when it's actually closed.

| non_test | session close message and actual session closing is out of sync sometimes when people are on spotty connections they have issues with timby saying your session is closed but then they still send a request and update their session before it closes these need to be in sync so the your session is closed message is only sent when it s actually closed | 0 |
166,683 | 20,719,828,321 | IssuesEvent | 2022-03-13 07:47:30 | mheob/itsb-web | https://api.github.com/repos/mheob/itsb-web | opened | prettier-stylelint-0.4.2.tgz: 3 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prettier-stylelint-0.4.2.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-28469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-2.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | multiple | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28469</summary>
### Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33623</summary>
### Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p>
<p>Trim newlines from the start and/or end of a string</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/trim-newlines/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- meow-3.7.0.tgz
- :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.
<p>Publish Date: 2021-05-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p>
<p>Release Date: 2021-05-28</p>
<p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-23382</summary>
### Vulnerable Libraries - <b>postcss-6.0.23.tgz</b>, <b>postcss-5.2.18.tgz</b></p>
<p>
### <b>postcss-6.0.23.tgz</b></p>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- :x: **postcss-6.0.23.tgz** (Vulnerable Library)
### <b>postcss-5.2.18.tgz</b></p>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- postcss-less-1.1.5.tgz
- :x: **postcss-5.2.18.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"2.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;micromatch:2.3.11;parse-glob:3.0.4;glob-base:0.3.0;glob-parent:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss-less:1.1.5;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | True | prettier-stylelint-0.4.2.tgz: 3 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prettier-stylelint-0.4.2.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-28469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-2.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | multiple | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28469</summary>
### Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33623</summary>
### Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p>
<p>Trim newlines from the start and/or end of a string</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/trim-newlines/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- meow-3.7.0.tgz
- :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.
<p>Publish Date: 2021-05-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p>
<p>Release Date: 2021-05-28</p>
<p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-23382</summary>
### Vulnerable Libraries - <b>postcss-6.0.23.tgz</b>, <b>postcss-5.2.18.tgz</b></p>
<p>
### <b>postcss-6.0.23.tgz</b></p>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- :x: **postcss-6.0.23.tgz** (Vulnerable Library)
### <b>postcss-5.2.18.tgz</b></p>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- prettier-stylelint-0.4.2.tgz (Root Library)
- stylelint-8.4.0.tgz
- postcss-less-1.1.5.tgz
- :x: **postcss-5.2.18.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"2.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;micromatch:2.3.11;parse-glob:3.0.4;glob-base:0.3.0;glob-parent:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss-less:1.1.5;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | non_test | prettier stylelint tgz vulnerabilities highest severity is vulnerable library prettier stylelint tgz path to dependency file frontend package json path to vulnerable library frontend node modules postcss less node modules postcss package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high glob parent tgz transitive n a high trim newlines tgz transitive n a medium multiple transitive n a details cve vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules glob parent package json dependency hierarchy prettier stylelint tgz root library stylelint tgz micromatch tgz parse glob tgz glob base tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource cve vulnerable library trim newlines tgz trim newlines from the start and or end of a string library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules trim newlines package json dependency hierarchy prettier stylelint tgz root library meow tgz x trim newlines tgz vulnerable library found in head commit a href found in base branch master vulnerability details the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim newlines step up your open source security game with whitesource cve vulnerable libraries postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules postcss package json dependency hierarchy prettier stylelint tgz root library stylelint tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules postcss less node modules postcss package json dependency hierarchy prettier stylelint tgz root library stylelint tgz postcss less tgz x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss step up your open source security game with whitesource istransitivedependency true dependencytree prettier stylelint stylelint micromatch parse glob glob base glob parent isminimumfixversionavailable true minimumfixversion glob parent isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator vulnerabilityurl istransitivedependency true dependencytree prettier stylelint meow trim newlines isminimumfixversionavailable true minimumfixversion trim newlines isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method vulnerabilityurl istransitivedependency true dependencytree prettier stylelint stylelint postcss isminimumfixversionavailable true minimumfixversion postcss isbinary false packagetype javascript node js packagename postcss packageversion packagefilepaths istransitivedependency true dependencytree prettier stylelint stylelint postcss less postcss isminimumfixversionavailable true minimumfixversion postcss isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl vulnerabilityurl | 0 |
197,502 | 22,595,470,453 | IssuesEvent | 2022-06-29 02:11:06 | interserver/mailbaby-mail-api | https://api.github.com/repos/interserver/mailbaby-mail-api | closed | WS-2021-0461 (Medium) detected in swagger-ui-3.45.1.js, swagger-ui-bundle-3.45.1.js - autoclosed | security vulnerability | ## WS-2021-0461 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>swagger-ui-3.45.1.js</b>, <b>swagger-ui-bundle-3.45.1.js</b></p></summary>
<p>
<details><summary><b>swagger-ui-3.45.1.js</b></p></summary>
<p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js</a></p>
<p>Path to vulnerable library: /public/doc/swagger-ui.js</p>
<p>
Dependency Hierarchy:
- :x: **swagger-ui-3.45.1.js** (Vulnerable Library)
</details>
<details><summary><b>swagger-ui-bundle-3.45.1.js</b></p></summary>
<p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js</a></p>
<p>Path to dependency file: /public/doc/index.html</p>
<p>Path to vulnerable library: /public/doc/./swagger-ui-bundle.js,/public/doc/swagger-ui-bundle.js</p>
<p>
Dependency Hierarchy:
- :x: **swagger-ui-bundle-3.45.1.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
SwaggerUI supports displaying remote OpenAPI definitions through the ?url parameter. This enables robust demonstration capabilities on sites like petstore.swagger.io, editor.swagger.io, and similar sites, where users often want to see what their OpenAPI definitions would look like rendered.
However, this functionality may pose a risk for users who host their own SwaggerUI instances. In particular, including remote OpenAPI definitions opens a vector for phishing attacks by abusing the trusted names/domains of self-hosted instances.
Resolution:
We've made the decision to disable query parameters (#4872) by default starting with SwaggerUI version 4.1.3. Please update to this version when it becomes available (ETA: 2021 December). Users will still be able to be re-enable the options at their discretion. We'll continue to enable query parameters on the Swagger demo sites.
<p>Publish Date: 2021-12-09
<p>URL: <a href=https://github.com/swagger-api/swagger-ui/commit/01a3e55960f864a0acf6a8d06e5ddaf6776a7f76>WS-2021-0461</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-qrmm-w75w-3wpx">https://github.com/advisories/GHSA-qrmm-w75w-3wpx</a></p>
<p>Release Date: 2021-12-09</p>
<p>Fix Resolution: swagger-ui - 4.1.3;swagger-ui-dist - 4.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0461 (Medium) detected in swagger-ui-3.45.1.js, swagger-ui-bundle-3.45.1.js - autoclosed - ## WS-2021-0461 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>swagger-ui-3.45.1.js</b>, <b>swagger-ui-bundle-3.45.1.js</b></p></summary>
<p>
<details><summary><b>swagger-ui-3.45.1.js</b></p></summary>
<p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js</a></p>
<p>Path to vulnerable library: /public/doc/swagger-ui.js</p>
<p>
Dependency Hierarchy:
- :x: **swagger-ui-3.45.1.js** (Vulnerable Library)
</details>
<details><summary><b>swagger-ui-bundle-3.45.1.js</b></p></summary>
<p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js</a></p>
<p>Path to dependency file: /public/doc/index.html</p>
<p>Path to vulnerable library: /public/doc/./swagger-ui-bundle.js,/public/doc/swagger-ui-bundle.js</p>
<p>
Dependency Hierarchy:
- :x: **swagger-ui-bundle-3.45.1.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
SwaggerUI supports displaying remote OpenAPI definitions through the ?url parameter. This enables robust demonstration capabilities on sites like petstore.swagger.io, editor.swagger.io, and similar sites, where users often want to see what their OpenAPI definitions would look like rendered.
However, this functionality may pose a risk for users who host their own SwaggerUI instances. In particular, including remote OpenAPI definitions opens a vector for phishing attacks by abusing the trusted names/domains of self-hosted instances.
Resolution:
We've made the decision to disable query parameters (#4872) by default starting with SwaggerUI version 4.1.3. Please update to this version when it becomes available (ETA: 2021 December). Users will still be able to be re-enable the options at their discretion. We'll continue to enable query parameters on the Swagger demo sites.
<p>Publish Date: 2021-12-09
<p>URL: <a href=https://github.com/swagger-api/swagger-ui/commit/01a3e55960f864a0acf6a8d06e5ddaf6776a7f76>WS-2021-0461</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-qrmm-w75w-3wpx">https://github.com/advisories/GHSA-qrmm-w75w-3wpx</a></p>
<p>Release Date: 2021-12-09</p>
<p>Fix Resolution: swagger-ui - 4.1.3;swagger-ui-dist - 4.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in swagger ui js swagger ui bundle js autoclosed ws medium severity vulnerability vulnerable libraries swagger ui js swagger ui bundle js swagger ui js swagger ui is a dependency free collection of html javascript and css assets that dynamically generate beautiful documentation from a swagger compliant api library home page a href path to vulnerable library public doc swagger ui js dependency hierarchy x swagger ui js vulnerable library swagger ui bundle js swagger ui is a dependency free collection of html javascript and css assets that dynamically generate beautiful documentation from a swagger compliant api library home page a href path to dependency file public doc index html path to vulnerable library public doc swagger ui bundle js public doc swagger ui bundle js dependency hierarchy x swagger ui bundle js vulnerable library found in head commit a href found in base branch master vulnerability details swaggerui supports displaying remote openapi definitions through the url parameter this enables robust demonstration capabilities on sites like petstore swagger io editor swagger io and similar sites where users often want to see what their openapi definitions would look like rendered however this functionality may pose a risk for users who host their own swaggerui instances in particular including remote openapi definitions opens a vector for phishing attacks by abusing the trusted names domains of self hosted instances resolution we ve made the decision to disable query parameters by default starting with swaggerui version please update to this version when it becomes available eta december users will still be able to be re enable the options at their discretion we ll continue to enable query parameters on the swagger demo sites publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution swagger ui swagger ui dist step up your open source security game with mend | 0 |
218,847 | 17,026,567,590 | IssuesEvent | 2021-07-03 16:46:24 | ZinedineMess/Billed-app-FR | https://api.github.com/repos/ZinedineMess/Billed-app-FR | closed | TU composant container/Bills | tests | - [x] couvrir tous les "statements" sauf les appels au back-end firebase (ils sont signalés en commentaire dans le code) : c'est simple, il faut que [le rapport de couverture du fichier container/Bills](http://127.0.0.1:8080/coverage/lcov-report/containers/Bills.js.html) soit vert.
Cela devrait permettre d'obtenir un taux de couverture aux alentours de 80% dans la colonne "statements". | 1.0 | TU composant container/Bills - - [x] couvrir tous les "statements" sauf les appels au back-end firebase (ils sont signalés en commentaire dans le code) : c'est simple, il faut que [le rapport de couverture du fichier container/Bills](http://127.0.0.1:8080/coverage/lcov-report/containers/Bills.js.html) soit vert.
Cela devrait permettre d'obtenir un taux de couverture aux alentours de 80% dans la colonne "statements". | test | tu composant container bills couvrir tous les statements sauf les appels au back end firebase ils sont signalés en commentaire dans le code c est simple il faut que soit vert cela devrait permettre d obtenir un taux de couverture aux alentours de dans la colonne statements | 1 |
290,797 | 25,096,497,782 | IssuesEvent | 2022-11-08 10:31:24 | Tencent/spring-cloud-tencent | https://api.github.com/repos/Tencent/spring-cloud-tencent | closed | 【需求认领】提升 spring-cloud-tencent-gateway-plugin 模块的单元测试覆盖率 | good first issue unit test | - 目标提升到80%以上
<img width="752" alt="image" src="https://user-images.githubusercontent.com/24446200/199162123-df7bb42d-e152-4bcd-b116-ae51aceb76fc.png">
- 如何获取单侧覆盖率:
1. 项目根目录下运行`mvn test`
2. 项目根目录下用浏览器打开`target/site/jacoco/index.html`即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充
- 额外提供
需要额外提供更新后的单测覆盖率截图到 PR 下方 | 1.0 | 【需求认领】提升 spring-cloud-tencent-gateway-plugin 模块的单元测试覆盖率 - - 目标提升到80%以上
<img width="752" alt="image" src="https://user-images.githubusercontent.com/24446200/199162123-df7bb42d-e152-4bcd-b116-ae51aceb76fc.png">
- 如何获取单侧覆盖率:
1. 项目根目录下运行`mvn test`
2. 项目根目录下用浏览器打开`target/site/jacoco/index.html`即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充
- 额外提供
需要额外提供更新后的单测覆盖率截图到 PR 下方 | test | 【需求认领】提升 spring cloud tencent gateway plugin 模块的单元测试覆盖率 以上 img width alt image src 如何获取单侧覆盖率: 项目根目录下运行 mvn test 项目根目录下用浏览器打开 target site jacoco index html 即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充 额外提供 需要额外提供更新后的单测覆盖率截图到 pr 下方 | 1 |
252,464 | 21,579,491,148 | IssuesEvent | 2022-05-02 17:06:58 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | Editing Secret object in Rancher UI corrupting binary data | kind/bug [zube]: To Test internal priority/1 team/area1 | Internal reference: SURE-4281
Reported in 2.6.3
The binary data in the Secret object getting corrupted after editing in the Rancher UI. The issue we have come across now is that if we use the webui to update an existing secret that contains a key/value pair with binary data and for example add a new key/value pair and save that the binary data becomes malformed.
Repro steps:
- Use a Secret manifest file. Create the Secret object with kubectl
- go to Rancher UI and edit the secret. I added some test key/value in the UI
- `kubectl get secret secret-name -o yaml` is shows different values for binary data.
Before & after show control characters appearing via some conversion.
Before:
```
apiVersion: v1
data:
amq-client.ks: MIIURQIBAzCCE/4GCSqGSIb3DQEHAaCCE+8EghPrMIIT5zCCCsMGCSqGSIb3DQEHAaCCCrQEggqwMIIKrDCCBU4GCyqGSIb3DQEMCgECoIIE+zCCBPcwKQYKKoZIhvcNAQwBAzAbBBQvb+FEMTTSYRXDivpl53xIWp
```
After:
```
apiVersion: v1
data:
amq-client.ks: MO+/vRRFAgEDMO+/vRPvv70GCSrvv71I77+977+9DQEHAe+/ve+/vRPvv70E77+9E++/vTDvv70T77+9MO+/vQrvv70GCSrvv71I77+977+9DQEHAe+/ve+/vQrvv70E77+9Cu+/vTDvv70K77+9MO+/vQVOBgsq77+9SO+/ve+/vQ0BDAoBAu+/ve+/vQTvv70w77+9BO+/vTApBgoq77+9SO+/ve+/
```
Workaround:
Is workaround available and implemented? yes
What is the workaround: using the `kubectl` command to edit the secrets.
Original work (for 2.4.6): https://github.com/rancher/rancher/issues/27439 | 1.0 | Editing Secret object in Rancher UI corrupting binary data - Internal reference: SURE-4281
Reported in 2.6.3
The binary data in the Secret object getting corrupted after editing in the Rancher UI. The issue we have come across now is that if we use the webui to update an existing secret that contains a key/value pair with binary data and for example add a new key/value pair and save that the binary data becomes malformed.
Repro steps:
- Use a Secret manifest file. Create the Secret object with kubectl
- go to Rancher UI and edit the secret. I added some test key/value in the UI
- `kubectl get secret secret-name -o yaml` is shows different values for binary data.
Before & after show control characters appearing via some conversion.
Before:
```
apiVersion: v1
data:
amq-client.ks: MIIURQIBAzCCE/4GCSqGSIb3DQEHAaCCE+8EghPrMIIT5zCCCsMGCSqGSIb3DQEHAaCCCrQEggqwMIIKrDCCBU4GCyqGSIb3DQEMCgECoIIE+zCCBPcwKQYKKoZIhvcNAQwBAzAbBBQvb+FEMTTSYRXDivpl53xIWp
```
After:
```
apiVersion: v1
data:
amq-client.ks: MO+/vRRFAgEDMO+/vRPvv70GCSrvv71I77+977+9DQEHAe+/ve+/vRPvv70E77+9E++/vTDvv70T77+9MO+/vQrvv70GCSrvv71I77+977+9DQEHAe+/ve+/vQrvv70E77+9Cu+/vTDvv70K77+9MO+/vQVOBgsq77+9SO+/ve+/vQ0BDAoBAu+/ve+/vQTvv70w77+9BO+/vTApBgoq77+9SO+/ve+/
```
Workaround:
Is workaround available and implemented? yes
What is the workaround: using the `kubectl` command to edit the secrets.
Original work (for 2.4.6): https://github.com/rancher/rancher/issues/27439 | test | editing secret object in rancher ui corrupting binary data internal reference sure reported in the binary data in the secret object getting corrupted after editing in the rancher ui the issue we have come across now is that if we use the webui to update an existing secret that contains a key value pair with binary data and for example add a new key value pair and save that the binary data becomes malformed repro steps use a secret manifest file create the secret object with kubectl go to rancher ui and edit the secret i added some test key value in the ui kubectl get secret secret name o yaml is shows different values for binary data before after show control characters appearing via some conversion before apiversion data amq client ks miiurqibazcce zccbpcwkqykkozihvcnaqwbazabbbqvb after apiversion data amq client ks mo vrrfagedmo ve ve ve ve ve workaround is workaround available and implemented yes what is the workaround using the kubectl command to edit the secrets original work for | 1 |
287,975 | 31,856,865,337 | IssuesEvent | 2023-09-15 08:07:09 | nidhi7598/linux-4.19.72_CVE-2022-3564 | https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564 | closed | CVE-2023-1390 (High) detected in linuxlinux-4.19.294 - autoclosed | Mend: dependency security vulnerability | ## CVE-2023-1390 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A remote denial of service vulnerability was found in the Linux kernel’s TIPC kernel module. The while loop in tipc_link_xmit() hits an unknown state while attempting to parse SKBs, which are not in the queue. Sending two small UDP packets to a system with a UDP bearer results in the CPU utilization for the system to instantly spike to 100%, causing a denial of service condition.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1390>CVE-2023-1390</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1390">https://www.linuxkernelcves.com/cves/CVE-2023-1390</a></p>
<p>Release Date: 2023-03-16</p>
<p>Fix Resolution: v4.9.253,v4.14.217,v4.19.170,v5.4.92,v5.10.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-1390 (High) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2023-1390 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A remote denial of service vulnerability was found in the Linux kernel’s TIPC kernel module. The while loop in tipc_link_xmit() hits an unknown state while attempting to parse SKBs, which are not in the queue. Sending two small UDP packets to a system with a UDP bearer results in the CPU utilization for the system to instantly spike to 100%, causing a denial of service condition.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1390>CVE-2023-1390</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1390">https://www.linuxkernelcves.com/cves/CVE-2023-1390</a></p>
<p>Release Date: 2023-03-16</p>
<p>Fix Resolution: v4.9.253,v4.14.217,v4.19.170,v5.4.92,v5.10.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a remote denial of service vulnerability was found in the linux kernel’s tipc kernel module the while loop in tipc link xmit hits an unknown state while attempting to parse skbs which are not in the queue sending two small udp packets to a system with a udp bearer results in the cpu utilization for the system to instantly spike to causing a denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
239,186 | 19,827,802,176 | IssuesEvent | 2022-01-20 08:50:30 | CoffeeKumazaki/arXiv | https://api.github.com/repos/CoffeeKumazaki/arXiv | closed | AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED) | Simulation Testing | AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED)<br>
https://ift.tt/3nUFyja<br>
<p>As self-driving systems become better, simulating scenarios where the autonomy stack may fail becomes more important. Traditionally, those scenarios are generated for a few scenes with respect to the planning module that takes ground-truth actor states as input. This does not scale and cannot identify all possible autonomy failures, such as perception failures due to occlusion. In this paper, we propose AdvSim, an adversarial framework to generate safety-critical scenarios for any LiDAR-based autonomy system. Given an initial traffic scenario, AdvSim modifies the actors' trajectories in a physically plausible manner and updates the LiDAR sensor data to match the perturbed world. Importantly, by simulating directly from sensor data, we obtain adversarial scenarios that are safety-critical for the full autonomy stack. Our experiments show that our approach is general and can identify thousands of semantically meaningful safety-critical scenarios for a wide range of modern self-driving systems. Furthermore, we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by AdvSim.</p>
<br>
<br>
via cs.RO updates on arXiv.org http://arxiv.org/<br> | 1.0 | AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED) - AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED)<br>
https://ift.tt/3nUFyja<br>
<p>As self-driving systems become better, simulating scenarios where the autonomy stack may fail becomes more important. Traditionally, those scenarios are generated for a few scenes with respect to the planning module that takes ground-truth actor states as input. This does not scale and cannot identify all possible autonomy failures, such as perception failures due to occlusion. In this paper, we propose AdvSim, an adversarial framework to generate safety-critical scenarios for any LiDAR-based autonomy system. Given an initial traffic scenario, AdvSim modifies the actors' trajectories in a physically plausible manner and updates the LiDAR sensor data to match the perturbed world. Importantly, by simulating directly from sensor data, we obtain adversarial scenarios that are safety-critical for the full autonomy stack. Our experiments show that our approach is general and can identify thousands of semantically meaningful safety-critical scenarios for a wide range of modern self-driving systems. Furthermore, we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by AdvSim.</p>
<br>
<br>
via cs.RO updates on arXiv.org http://arxiv.org/<br> | test | advsim generating safety critical scenarios for self driving vehicles arxiv updated advsim generating safety critical scenarios for self driving vehicles arxiv updated as self driving systems become better simulating scenarios where the autonomy stack may fail becomes more important traditionally those scenarios are generated for a few scenes with respect to the planning module that takes ground truth actor states as input this does not scale and cannot identify all possible autonomy failures such as perception failures due to occlusion in this paper we propose advsim an adversarial framework to generate safety critical scenarios for any lidar based autonomy system given an initial traffic scenario advsim modifies the actors trajectories in a physically plausible manner and updates the lidar sensor data to match the perturbed world importantly by simulating directly from sensor data we obtain adversarial scenarios that are safety critical for the full autonomy stack our experiments show that our approach is general and can identify thousands of semantically meaningful safety critical scenarios for a wide range of modern self driving systems furthermore we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by advsim via cs ro updates on arxiv org | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.