Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
76,892
7,548,573,543
IssuesEvent
2018-04-18 11:42:20
rucio/rucio
https://api.github.com/repos/rucio/rucio
closed
Work on the demo
Testing enhancement patch
Motivation ---------- Demo is empty Modification ------------ Create RSEs, DIDs...
1.0
Work on the demo - Motivation ---------- Demo is empty Modification ------------ Create RSEs, DIDs...
test
work on the demo motivation demo is empty modification create rses dids
1
61,627
8,531,489,474
IssuesEvent
2018-11-04 12:54:58
49thSecurityDivision/lab-documentation
https://api.github.com/repos/49thSecurityDivision/lab-documentation
opened
Snipe-IT submit new assest
documentation lab_minions
**Describe what prompted this request** The process of how we onboard new items when we buy them. **Describe the solution you'd like** Show all steps for adding new items for minions. **Who is the target user for this?** Minions **Will it need to be maintained?** No, unless snipe-it changes **Who should have access?** Lab minions **Are these docs testable? (not all documentation will be testable)** No **Will there be sensitive information in these docs?** No **Additional context** Since Snipe-IT is our asset management system, we need to build a workflow around what happens when we get/purchase new item(s).
1.0
Snipe-IT submit new assest - **Describe what prompted this request** The process of how we onboard new items when we buy them. **Describe the solution you'd like** Show all steps for adding new items for minions. **Who is the target user for this?** Minions **Will it need to be maintained?** No, unless snipe-it changes **Who should have access?** Lab minions **Are these docs testable? (not all documentation will be testable)** No **Will there be sensitive information in these docs?** No **Additional context** Since Snipe-IT is our asset management system, we need to build a workflow around what happens when we get/purchase new item(s).
non_test
snipe it submit new assest describe what prompted this request the process of how we onboard new items when we buy them describe the solution you d like show all steps for adding new items for minions who is the target user for this minions will it need to be maintained no unless snipe it changes who should have access lab minions are these docs testable not all documentation will be testable no will there be sensitive information in these docs no additional context since snipe it is our asset management system we need to build a workflow around what happens when we get purchase new item s
0
133,857
18,358,481,922
IssuesEvent
2021-10-08 22:24:36
GabDug/MaterialIconsLibrary
https://api.github.com/repos/GabDug/MaterialIconsLibrary
opened
CVE-2021-32804 (High) detected in tar-6.1.0.tgz
security vulnerability
## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p> <p>Path to dependency file: MaterialIconsLibrary/package.json</p> <p>Path to vulnerable library: MaterialIconsLibrary/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - cli-12.0.0.tgz (Root Library) - pacote-11.3.2.tgz - :x: **tar-6.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GabDug/MaterialIconsLibrary/commit/89aa5982db5f213f61df052ca5dc7f0732ce3a91">89aa5982db5f213f61df052ca5dc7f0732ce3a91</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32804 (High) detected in tar-6.1.0.tgz - ## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p> <p>Path to dependency file: MaterialIconsLibrary/package.json</p> <p>Path to vulnerable library: MaterialIconsLibrary/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - cli-12.0.0.tgz (Root Library) - pacote-11.3.2.tgz - :x: **tar-6.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GabDug/MaterialIconsLibrary/commit/89aa5982db5f213f61df052ca5dc7f0732ce3a91">89aa5982db5f213f61df052ca5dc7f0732ce3a91</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file materialiconslibrary package json path to vulnerable library materialiconslibrary node modules tar package json dependency hierarchy cli tgz root library pacote tgz x tar tgz vulnerable library found in head commit a href found in base branch main vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
134,175
10,885,485,490
IssuesEvent
2019-11-18 10:29:33
SatelliteQE/robottelo
https://api.github.com/repos/SatelliteQE/robottelo
closed
cli.test_user.UserWithCleanUpTestCase.test_positive_update_all_roles not isolated
6.3 6.4 Bug test-failure
besides that i don't really know what this test is supposed to really test (it look quite useless to me), it also lacks the isolation: It first lists all the roles and then creates an user and assigns the roles to it one by one. - However, some other role-related test runs in parralel, creating and managing its own custom role, including deleting it at the end. This causes the first test to discover this role during initial listing and sometimes assigning, however it fails to find it at the end during the assertion listing, since it no longer exists
1.0
cli.test_user.UserWithCleanUpTestCase.test_positive_update_all_roles not isolated - besides that i don't really know what this test is supposed to really test (it look quite useless to me), it also lacks the isolation: It first lists all the roles and then creates an user and assigns the roles to it one by one. - However, some other role-related test runs in parralel, creating and managing its own custom role, including deleting it at the end. This causes the first test to discover this role during initial listing and sometimes assigning, however it fails to find it at the end during the assertion listing, since it no longer exists
test
cli test user userwithcleanuptestcase test positive update all roles not isolated besides that i don t really know what this test is supposed to really test it look quite useless to me it also lacks the isolation it first lists all the roles and then creates an user and assigns the roles to it one by one however some other role related test runs in parralel creating and managing its own custom role including deleting it at the end this causes the first test to discover this role during initial listing and sometimes assigning however it fails to find it at the end during the assertion listing since it no longer exists
1
8,916
3,010,872,727
IssuesEvent
2015-07-28 15:13:52
brharp/hjckrrh
https://api.github.com/repos/brharp/hjckrrh
closed
Bootstrap Anchor issue
priority: high status: fixed - testing required type: bug
Platform configuration change needed - When in Home |Administration |Appearance | Settings in Anchors sections - need to to deselect the "fix anchor positions". Otherwise anchors don't work in bootstrap.
1.0
Bootstrap Anchor issue - Platform configuration change needed - When in Home |Administration |Appearance | Settings in Anchors sections - need to to deselect the "fix anchor positions". Otherwise anchors don't work in bootstrap.
test
bootstrap anchor issue platform configuration change needed when in home administration appearance settings in anchors sections need to to deselect the fix anchor positions otherwise anchors don t work in bootstrap
1
164,382
12,801,054,229
IssuesEvent
2020-07-02 18:20:37
syscoin/syscoin
https://api.github.com/repos/syscoin/syscoin
opened
New masternode commands causes crash when entered without argument
4.2-testnet
Issue most likely due to ported commands being non-standard to the bitcoin's new rpc format **Expected behavior** ``` $ syscoin-cli bls error code: -1 error message: bls Set of commands to execute BLS related actions. To get help on individual commands, use "help bls command". Arguments: 1. "command" (string, required) The command to execute Available commands: generate - Create a BLS secret/public key pair fromsecret - Parse a BLS secret key and return the secret/public key pair Result: null (json null) ``` ``` $ syscoin-cli help protx error code: -1 error message: protx Set of commands to execute ProTx related actions. To get help on individual commands, use "help protx command". Arguments: 1. "command" (string, required) The command to execute Available commands: register - Create and send ProTx to network register_fund - Fund, create and send ProTx to network register_prepare - Create an unsigned ProTx register_submit - Sign and submit a ProTx list - List ProTxs info - Return information about a ProTx update_service - Create and send ProUpServTx to network update_registrar - Create and send ProUpRegTx to network revoke - Create and send ProUpRevTx to network diff - Calculate a diff and a proof between two masternode lists Result: null (json null) ``` <!--- What behavior did you expect? --> **Actual behavior** ``` $ syscoin-cli bls error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached") Make sure the syscoind server is running and that you are connecting to the correct RPC port. ``` ``` $ syscoin-cli protx error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached") Make sure the syscoind server is running and that you are connecting to the correct RPC port. ``` <!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? --> **To reproduce** `syscoin-cli bls` `syscoin-cli masternode` `syscoin-cli protx`
1.0
New masternode commands causes crash when entered without argument - Issue most likely due to ported commands being non-standard to the bitcoin's new rpc format **Expected behavior** ``` $ syscoin-cli bls error code: -1 error message: bls Set of commands to execute BLS related actions. To get help on individual commands, use "help bls command". Arguments: 1. "command" (string, required) The command to execute Available commands: generate - Create a BLS secret/public key pair fromsecret - Parse a BLS secret key and return the secret/public key pair Result: null (json null) ``` ``` $ syscoin-cli help protx error code: -1 error message: protx Set of commands to execute ProTx related actions. To get help on individual commands, use "help protx command". Arguments: 1. "command" (string, required) The command to execute Available commands: register - Create and send ProTx to network register_fund - Fund, create and send ProTx to network register_prepare - Create an unsigned ProTx register_submit - Sign and submit a ProTx list - List ProTxs info - Return information about a ProTx update_service - Create and send ProUpServTx to network update_registrar - Create and send ProUpRegTx to network revoke - Create and send ProUpRevTx to network diff - Calculate a diff and a proof between two masternode lists Result: null (json null) ``` <!--- What behavior did you expect? --> **Actual behavior** ``` $ syscoin-cli bls error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached") Make sure the syscoind server is running and that you are connecting to the correct RPC port. ``` ``` $ syscoin-cli protx error: Could not connect to the server 127.0.0.1:18370 (error code 1 - "EOF reached") Make sure the syscoind server is running and that you are connecting to the correct RPC port. ``` <!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? --> **To reproduce** `syscoin-cli bls` `syscoin-cli masternode` `syscoin-cli protx`
test
new masternode commands causes crash when entered without argument issue most likely due to ported commands being non standard to the bitcoin s new rpc format expected behavior syscoin cli bls error code error message bls set of commands to execute bls related actions to get help on individual commands use help bls command arguments command string required the command to execute available commands generate create a bls secret public key pair fromsecret parse a bls secret key and return the secret public key pair result null json null syscoin cli help protx error code error message protx set of commands to execute protx related actions to get help on individual commands use help protx command arguments command string required the command to execute available commands register create and send protx to network register fund fund create and send protx to network register prepare create an unsigned protx register submit sign and submit a protx list list protxs info return information about a protx update service create and send proupservtx to network update registrar create and send proupregtx to network revoke create and send prouprevtx to network diff calculate a diff and a proof between two masternode lists result null json null actual behavior syscoin cli bls error could not connect to the server error code eof reached make sure the syscoind server is running and that you are connecting to the correct rpc port syscoin cli protx error could not connect to the server error code eof reached make sure the syscoind server is running and that you are connecting to the correct rpc port to reproduce syscoin cli bls syscoin cli masternode syscoin cli protx
1
188,063
6,767,976,918
IssuesEvent
2017-10-26 06:57:47
edenlabllc/ehealth.api
https://api.github.com/repos/edenlabllc/ehealth.api
closed
OTP SMS delivery metrics
epic/sms kind/user_story priority/medium status/wontfix
We should have a metrics for SMS delivery process * Succesful/unsuccessful SMS submissions stats * Undelivered SMS * SMS delivery latency - [ ] integration with life report to store counters - [ ] new metrics on datadog https://docs.google.com/spreadsheets/d/1X1gQEWQc02loG1OtNRZzzuN3NssLRoESIgpn-aRDMPQ/edit?usp=sharing
1.0
OTP SMS delivery metrics - We should have a metrics for SMS delivery process * Succesful/unsuccessful SMS submissions stats * Undelivered SMS * SMS delivery latency - [ ] integration with life report to store counters - [ ] new metrics on datadog https://docs.google.com/spreadsheets/d/1X1gQEWQc02loG1OtNRZzzuN3NssLRoESIgpn-aRDMPQ/edit?usp=sharing
non_test
otp sms delivery metrics we should have a metrics for sms delivery process succesful unsuccessful sms submissions stats undelivered sms sms delivery latency integration with life report to store counters new metrics on datadog
0
121,715
10,193,878,281
IssuesEvent
2019-08-12 14:34:01
sparkdesignsystem/spark-design-system
https://api.github.com/repos/sparkdesignsystem/spark-design-system
closed
Follow/successfully pass Manual Test Plan for Promo
scope: all pkgs status: PO approved type: testing
**User Story:** As Spark we want to follow the Manual Test plan for Promo so that we can confirm Manual Test plan is up to date. **AC:** - Successful complete Manual Test Plan for Promo
1.0
Follow/successfully pass Manual Test Plan for Promo - **User Story:** As Spark we want to follow the Manual Test plan for Promo so that we can confirm Manual Test plan is up to date. **AC:** - Successful complete Manual Test Plan for Promo
test
follow successfully pass manual test plan for promo user story as spark we want to follow the manual test plan for promo so that we can confirm manual test plan is up to date ac successful complete manual test plan for promo
1
6,655
3,039,344,505
IssuesEvent
2015-08-07 08:13:40
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
reference documentation "67.7 Use Two EntityManagers" should be extended
documentation
the reference documentation "67.7 Use Two EntityManagers" describe the declaration for a second EntityManagerFactory, but the usage is obmitted. usage examples should be added. for example with spring data jpa: ``` @Configuration @EnableJpaRepositories(value = "...dataaccess.customer.repository", entityManagerFactoryRef = "customerEntityManagerFactory") public class CustomerInitializer {} ``` ``` @Configuration @EnableJpaRepositories(value = "...dataaccess.order.repository", entityManagerFactoryRef = "orderEntityManagerFactory") public class OrderInitializer {} ``` thx Rico
1.0
reference documentation "67.7 Use Two EntityManagers" should be extended - the reference documentation "67.7 Use Two EntityManagers" describe the declaration for a second EntityManagerFactory, but the usage is obmitted. usage examples should be added. for example with spring data jpa: ``` @Configuration @EnableJpaRepositories(value = "...dataaccess.customer.repository", entityManagerFactoryRef = "customerEntityManagerFactory") public class CustomerInitializer {} ``` ``` @Configuration @EnableJpaRepositories(value = "...dataaccess.order.repository", entityManagerFactoryRef = "orderEntityManagerFactory") public class OrderInitializer {} ``` thx Rico
non_test
reference documentation use two entitymanagers should be extended the reference documentation use two entitymanagers describe the declaration for a second entitymanagerfactory but the usage is obmitted usage examples should be added for example with spring data jpa configuration enablejparepositories value dataaccess customer repository entitymanagerfactoryref customerentitymanagerfactory public class customerinitializer configuration enablejparepositories value dataaccess order repository entitymanagerfactoryref orderentitymanagerfactory public class orderinitializer thx rico
0
185,396
14,351,682,038
IssuesEvent
2020-11-30 02:03:54
CSU-Booking-Platform/application
https://api.github.com/repos/CSU-Booking-Platform/application
opened
Acceptance Tests for #34 Create Permissions
acceptance-test
### User story #34 ### Acceptance Test Checklist - [ ] If I am an administrator, when I try to create a permission, the system registers a new permission - [ ] If I am not an administrator, when I try to create a permission, the system does not create the new permission
1.0
Acceptance Tests for #34 Create Permissions - ### User story #34 ### Acceptance Test Checklist - [ ] If I am an administrator, when I try to create a permission, the system registers a new permission - [ ] If I am not an administrator, when I try to create a permission, the system does not create the new permission
test
acceptance tests for create permissions user story acceptance test checklist if i am an administrator when i try to create a permission the system registers a new permission if i am not an administrator when i try to create a permission the system does not create the new permission
1
310,022
26,694,645,965
IssuesEvent
2023-01-27 09:17:31
MetaMask/metamask-mobile
https://api.github.com/repos/MetaMask/metamask-mobile
opened
Improve test coverage of send flow
tests team-confirmations
Improve test coverage of send flow on mobile. This includes: 1. Writing more unit test coverage for different components used in send flow 2. Writing more e2e test for send screens: ![Screenshot 2023-01-27 at 2.44.06 PM.png](https://images.zenhubusercontent.com/61e70564eac52332e3564685/d0434c53-63a5-4aa0-8b9e-de7a1c6538a1)![Screenshot 2023-01-27 at 2.44.26 PM.png](https://images.zenhubusercontent.com/61e70564eac52332e3564685/d0af7fe4-9ca8-41a9-a5c1-60c976a1a6f4)
1.0
Improve test coverage of send flow - Improve test coverage of send flow on mobile. This includes: 1. Writing more unit test coverage for different components used in send flow 2. Writing more e2e test for send screens: ![Screenshot 2023-01-27 at 2.44.06 PM.png](https://images.zenhubusercontent.com/61e70564eac52332e3564685/d0434c53-63a5-4aa0-8b9e-de7a1c6538a1)![Screenshot 2023-01-27 at 2.44.26 PM.png](https://images.zenhubusercontent.com/61e70564eac52332e3564685/d0af7fe4-9ca8-41a9-a5c1-60c976a1a6f4)
test
improve test coverage of send flow improve test coverage of send flow on mobile this includes writing more unit test coverage for different components used in send flow writing more test for send screens
1
71,438
7,244,815,618
IssuesEvent
2018-02-14 16:11:02
eclipse/kura
https://api.github.com/repos/eclipse/kura
closed
[Testing] BLE Util test erratic
testing
BluetoothLeUtilTest.testKillCmd has a tendency to fail on CI. Remove it.
1.0
[Testing] BLE Util test erratic - BluetoothLeUtilTest.testKillCmd has a tendency to fail on CI. Remove it.
test
ble util test erratic bluetoothleutiltest testkillcmd has a tendency to fail on ci remove it
1
161,134
13,806,692,518
IssuesEvent
2020-10-11 18:44:19
codimd/server
https://api.github.com/repos/codimd/server
closed
Add note revisions documentation
documentation enhancement good first issue
After getting #195 and checking out features page, I had to notice, that we don't explain note revisions anywhere. ![note revisions](https://user-images.githubusercontent.com/8719867/66591137-8b84ce80-eb92-11e9-9473-38bd6624588e.png) We should add some documentation about them to the [features.md](https://github.com/codimd/server/blob/master/public/docs/features.md).
1.0
Add note revisions documentation - After getting #195 and checking out features page, I had to notice, that we don't explain note revisions anywhere. ![note revisions](https://user-images.githubusercontent.com/8719867/66591137-8b84ce80-eb92-11e9-9473-38bd6624588e.png) We should add some documentation about them to the [features.md](https://github.com/codimd/server/blob/master/public/docs/features.md).
non_test
add note revisions documentation after getting and checking out features page i had to notice that we don t explain note revisions anywhere we should add some documentation about them to the
0
226,055
17,937,153,821
IssuesEvent
2021-09-10 16:48:48
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
closed
//regen Spitting Errors in Console when attempting to execute command
Requires Testing
### Server Implementation Spigot ### Server Version 1.17.1 ### Describe the bug When attempting to use //regen after selecting an area, the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console. Some of the effects of the error are seen ingame, such as grass not generating on the surface or bedrock not generating. Cave generation will also be slightly different. Here is a paste of the error in console: [](https://pastebin.com/gyysY9Ft) ### To Reproduce 1. Select an area larger than 2000 blocks. Try to include the surface grass or bedrock for the clearest effect. 2. Run //regen ingame. 3. Errors should appear in console and the regenerated region will look different than the original vanilla terrain. ### Expected behaviour After using //regen I was expecting the area to generate as if it were being generated by minecraft for the first time, with identical structure and landform. ### Screenshots / Videos ![2021-09-10_11 42 44](https://user-images.githubusercontent.com/30448157/132888443-89e439c6-cb1c-4657-9280-713f0f7d0748.png) ![2021-09-10_11 43 03](https://user-images.githubusercontent.com/30448157/132888446-64e73723-5d5f-4974-8a86-db0f522514aa.png) ### Error log (if applicable) https://pastebin.com/gyysY9Ft ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/b12a1d36b6b544ffb7890ad1fa2b091e ### Fawe Version FastAsyncWorldEdit version 1.17-217;c16212e ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit-1.17/ and the issue still persists. ### Anything else? I've tested //regen in multiple areas of the overworld, and the issue remains consistent. I also am running an anticheat on the server (Vulcan), but I highly doubt this is the cause of the issue.
1.0
//regen Spitting Errors in Console when attempting to execute command - ### Server Implementation Spigot ### Server Version 1.17.1 ### Describe the bug When attempting to use //regen after selecting an area, the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console. Some of the effects of the error are seen ingame, such as grass not generating on the surface or bedrock not generating. Cave generation will also be slightly different. Here is a paste of the error in console: [](https://pastebin.com/gyysY9Ft) ### To Reproduce 1. Select an area larger than 2000 blocks. Try to include the surface grass or bedrock for the clearest effect. 2. Run //regen ingame. 3. Errors should appear in console and the regenerated region will look different than the original vanilla terrain. ### Expected behaviour After using //regen I was expecting the area to generate as if it were being generated by minecraft for the first time, with identical structure and landform. ### Screenshots / Videos ![2021-09-10_11 42 44](https://user-images.githubusercontent.com/30448157/132888443-89e439c6-cb1c-4657-9280-713f0f7d0748.png) ![2021-09-10_11 43 03](https://user-images.githubusercontent.com/30448157/132888446-64e73723-5d5f-4974-8a86-db0f522514aa.png) ### Error log (if applicable) https://pastebin.com/gyysY9Ft ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/b12a1d36b6b544ffb7890ad1fa2b091e ### Fawe Version FastAsyncWorldEdit version 1.17-217;c16212e ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit-1.17/ and the issue still persists. ### Anything else? I've tested //regen in multiple areas of the overworld, and the issue remains consistent. I also am running an anticheat on the server (Vulcan), but I highly doubt this is the cause of the issue.
test
regen spitting errors in console when attempting to execute command server implementation spigot server version describe the bug when attempting to use regen after selecting an area the region regenerates incorrectly compared to vanilla generation and a series of errors appears in console some of the effects of the error are seen ingame such as grass not generating on the surface or bedrock not generating cave generation will also be slightly different here is a paste of the error in console to reproduce select an area larger than blocks try to include the surface grass or bedrock for the clearest effect run regen ingame errors should appear in console and the regenerated region will look different than the original vanilla terrain expected behaviour after using regen i was expecting the area to generate as if it were being generated by minecraft for the first time with identical structure and landform screenshots videos error log if applicable fawe debugpaste fawe version fastasyncworldedit version checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else i ve tested regen in multiple areas of the overworld and the issue remains consistent i also am running an anticheat on the server vulcan but i highly doubt this is the cause of the issue
1
234,833
19,270,882,863
IssuesEvent
2021-12-10 05:09:06
cse110-fa21-group2/cse110-fa21-group2
https://api.github.com/repos/cse110-fa21-group2/cse110-fa21-group2
closed
Laurence's Testing Assignments
testing
#Functions to Test - function clearSortingandFiltering - async function populateExplore() - function initializeButtons() - createCardsFromData(items, container) - function initializeLocalStorage()
1.0
Laurence's Testing Assignments - #Functions to Test - function clearSortingandFiltering - async function populateExplore() - function initializeButtons() - createCardsFromData(items, container) - function initializeLocalStorage()
test
laurence s testing assignments functions to test function clearsortingandfiltering async function populateexplore function initializebuttons createcardsfromdata items container function initializelocalstorage
1
265,177
23,149,609,072
IssuesEvent
2022-07-29 06:54:07
cnosdb/cnosdb
https://api.github.com/repos/cnosdb/cnosdb
closed
[Bug]parallel kvcore test will panic
type/bug test kv store
**Describe the bug** parallel kvcore test will panic,we add serial for each test,the test will pass **To Reproduce** cargo test --package tskv --lib kvcore::test **Expected behavior** should parallel **Additional context** none
1.0
[Bug]parallel kvcore test will panic - **Describe the bug** parallel kvcore test will panic,we add serial for each test,the test will pass **To Reproduce** cargo test --package tskv --lib kvcore::test **Expected behavior** should parallel **Additional context** none
test
parallel kvcore test will panic describe the bug parallel kvcore test will panic,we add serial for each test,the test will pass to reproduce cargo test package tskv lib kvcore test expected behavior should parallel additional context none
1
148,000
19,526,266,752
IssuesEvent
2021-12-30 08:25:55
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2020-25672 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2020-25672 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak vulnerability was found in Linux kernel in llcp_sock_connect <p>Publish Date: 2021-05-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25672 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2020-25672 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak vulnerability was found in Linux kernel in llcp_sock_connect <p>Publish Date: 2021-05-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25672>CVE-2020-25672</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net nfc llcp sock c net nfc llcp sock c vulnerability details a memory leak vulnerability was found in linux kernel in llcp sock connect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
244,967
20,735,263,055
IssuesEvent
2022-03-14 13:12:24
Azure/azure-sdk-for-java
https://api.github.com/repos/Azure/azure-sdk-for-java
closed
Increase Unit Test Coverage for Azure Storage Blob
Storage Client help wanted test-enhancement
Add additional unit tests to get closer to 90% coverage. https://azuresdkartifacts.blob.core.windows.net/azure-sdk-for-java/test-coverage/azure-storage-blob/index.html#dn-c
1.0
Increase Unit Test Coverage for Azure Storage Blob - Add additional unit tests to get closer to 90% coverage. https://azuresdkartifacts.blob.core.windows.net/azure-sdk-for-java/test-coverage/azure-storage-blob/index.html#dn-c
test
increase unit test coverage for azure storage blob add additional unit tests to get closer to coverage
1
285,438
24,666,612,473
IssuesEvent
2022-10-18 10:46:24
strimzi/strimzi-kafka-operator
https://api.github.com/repos/strimzi/strimzi-kafka-operator
closed
[ST] Add message count, kafka + zk + eo STS or deployment names into the `TestStorage`
enhancement System tests
**Current situation** Currently we are using `TestStorage` in most of cases, but this storage doesn't contains information about message count (for clients exchange), Kafka & ZK & EO `StatefulSet` or `Deployment` names. It would be much better to add those (and others, which are needed) to the `TestStorage`, so we would be able to remove all extra variables from the tests and use just the `TestStorage`. **Describe the solution you'd like** Best solution would be to add, how I mentioned, the all needed information to the `TestStorage` and customize the methods with it - that means, that for example `ClientUtils.waitForClientSuccess` could work just with `TestStorage` and not the multiple other parameters - which is sometimes confusing and it could cause some future issues.
1.0
[ST] Add message count, kafka + zk + eo STS or deployment names into the `TestStorage` - **Current situation** Currently we are using `TestStorage` in most of cases, but this storage doesn't contains information about message count (for clients exchange), Kafka & ZK & EO `StatefulSet` or `Deployment` names. It would be much better to add those (and others, which are needed) to the `TestStorage`, so we would be able to remove all extra variables from the tests and use just the `TestStorage`. **Describe the solution you'd like** Best solution would be to add, how I mentioned, the all needed information to the `TestStorage` and customize the methods with it - that means, that for example `ClientUtils.waitForClientSuccess` could work just with `TestStorage` and not the multiple other parameters - which is sometimes confusing and it could cause some future issues.
test
add message count kafka zk eo sts or deployment names into the teststorage current situation currently we are using teststorage in most of cases but this storage doesn t contains information about message count for clients exchange kafka zk eo statefulset or deployment names it would be much better to add those and others which are needed to the teststorage so we would be able to remove all extra variables from the tests and use just the teststorage describe the solution you d like best solution would be to add how i mentioned the all needed information to the teststorage and customize the methods with it that means that for example clientutils waitforclientsuccess could work just with teststorage and not the multiple other parameters which is sometimes confusing and it could cause some future issues
1
152,364
12,102,649,016
IssuesEvent
2020-04-20 17:02:34
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
cluster k8s >=1.15 with Canal/Calico do not deploy in 2.3.0 and 2.3.2
[zube]: To Test kind/bug-qa
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - Deploy cluster using k8s - 1.15, 1.16 and 1.17 - The cluster fails to come up - etcd container logs: ``` 2020-03-28 01:44:40.270970 I | embed: rejected connection from "ip:42920" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:46:53.878418 I | embed: rejected connection from "ip:41891" (error "EOF", ServerName "") 2020-03-28 01:47:05.883684 I | embed: rejected connection from "ip:42954" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:47:11.027567 I | embed: rejected connection from "ip:42958" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:47:16.171899 I | embed: rejected connection from "ip:42960" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") ``` **Expected Result:** cluster should be deployed successfully **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.3.0 to 2.3.2 - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do - Kubernetes version (use `kubectl version`): ``` 1.15, 1.16 and 1.17 ```
1.0
cluster k8s >=1.15 with Canal/Calico do not deploy in 2.3.0 and 2.3.2 - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - Deploy cluster using k8s - 1.15, 1.16 and 1.17 - The cluster fails to come up - etcd container logs: ``` 2020-03-28 01:44:40.270970 I | embed: rejected connection from "ip:42920" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:46:53.878418 I | embed: rejected connection from "ip:41891" (error "EOF", ServerName "") 2020-03-28 01:47:05.883684 I | embed: rejected connection from "ip:42954" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:47:11.027567 I | embed: rejected connection from "ip:42958" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") 2020-03-28 01:47:16.171899 I | embed: rejected connection from "ip:42960" (error "tls: failed to verify client's certificate: x509: certificate signed by unknown authority (possibly because of \"crypto/rsa: verification error\" while trying to verify candidate authority certificate \"kube-ca\")", ServerName "") ``` **Expected Result:** cluster should be deployed successfully **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.3.0 to 2.3.2 - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do - Kubernetes version (use `kubectl version`): ``` 1.15, 1.16 and 1.17 ```
test
cluster with canal calico do not deploy in and what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible deploy cluster using and the cluster fails to come up etcd container logs i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error eof servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername i embed rejected connection from ip error tls failed to verify client s certificate certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate kube ca servername expected result cluster should be deployed successfully environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui to installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported rke do kubernetes version use kubectl version and
1
318,805
27,321,752,661
IssuesEvent
2023-02-24 20:32:08
celestiaorg/celestia-app
https://api.github.com/repos/celestiaorg/celestia-app
closed
`nmt_wrapper_test.go` depends on magic constant
testing
## Context https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper_test.go#L98 ## Problem Based on [these lines](https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper.go#L75-L80) I expect the entire share (i.e. 512 bytes) to be pushed to the erasured NMT. `appconsts.ContinuationSparseShareContentSize` + 1 may not always == the share size - the len of the namespace
1.0
`nmt_wrapper_test.go` depends on magic constant - ## Context https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper_test.go#L98 ## Problem Based on [these lines](https://github.com/celestiaorg/celestia-app/blob/e6bbc302194376dbfd909cfabbf82e16e8fd5e7f/pkg/wrapper/nmt_wrapper.go#L75-L80) I expect the entire share (i.e. 512 bytes) to be pushed to the erasured NMT. `appconsts.ContinuationSparseShareContentSize` + 1 may not always == the share size - the len of the namespace
test
nmt wrapper test go depends on magic constant context problem based on i expect the entire share i e bytes to be pushed to the erasured nmt appconsts continuationsparsesharecontentsize may not always the share size the len of the namespace
1
229,713
25,362,329,292
IssuesEvent
2022-11-21 01:08:08
Killy85/game_ai_trainer
https://api.github.com/repos/Killy85/game_ai_trainer
opened
CVE-2022-41896 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2022-41896 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. If `ThreadUnsafeUnigramCandidateSampler` is given input `filterbank_channel_count` greater than the allowed max size, TensorFlow will crash. We have patched the issue in GitHub commit 39ec7eaf1428e90c37787e5b3fbd68ebd3c48860. The fix will be included in TensorFlow 2.11. We will also cherrypick this commit on TensorFlow 2.10.1, 2.9.3, and TensorFlow 2.8.4, as these are also affected and still in supported range. <p>Publish Date: 2022-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41896>CVE-2022-41896</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-41896">https://www.cve.org/CVERecord?id=CVE-2022-41896</a></p> <p>Release Date: 2022-11-18</p> <p>Fix Resolution: tensorflow - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-cpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-gpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-41896 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2022-41896 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. If `ThreadUnsafeUnigramCandidateSampler` is given input `filterbank_channel_count` greater than the allowed max size, TensorFlow will crash. We have patched the issue in GitHub commit 39ec7eaf1428e90c37787e5b3fbd68ebd3c48860. The fix will be included in TensorFlow 2.11. We will also cherrypick this commit on TensorFlow 2.10.1, 2.9.3, and TensorFlow 2.8.4, as these are also affected and still in supported range. <p>Publish Date: 2022-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41896>CVE-2022-41896</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-41896">https://www.cve.org/CVERecord?id=CVE-2022-41896</a></p> <p>Release Date: 2022-11-18</p> <p>Fix Resolution: tensorflow - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-cpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0, tensorflow-gpu - 2.8.4, 2.9.3, 2.10.1, 2.11.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning if threadunsafeunigramcandidatesampler is given input filterbank channel count greater than the allowed max size tensorflow will crash we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
0
294,786
25,404,618,572
IssuesEvent
2022-11-22 14:29:45
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
SC_Softmx_JitAot_Linux hang
comp:jit test failure
https://openj9-jenkins.osuosl.org/job/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324 SC_Softmx_JitAot_Linux_1 `-Xcompressedrefs -Xjit -Xgcpolicy:gencon ` https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324/system_test_output.tar.gz ``` STF 01:15:28.337 - **FAILED** Process jvm4 has timed out STF 01:15:28.337 - Collecting dumps for: jvm4 ``` There are core/javacore/etc. diagnostic files. The javacore files show this state: ``` 2LKREGMON JIT-QueueSlotMonitor-1 lock (0x0000FFFF94003300): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "load-1" (J9VMThread:0x000000000057C200) 3LKWAITNOTIFY "load-2" (J9VMThread:0x000000000057E700) 3LKWAITNOTIFY "load-3" (J9VMThread:0x0000000000580B00) 3LKWAITNOTIFY "load-4" (J9VMThread:0x0000000000583000) 3LKWAITNOTIFY "load-5" (J9VMThread:0x0000000000585500) 3LKWAITNOTIFY "load-6" (J9VMThread:0x0000000000587A00) 2LKREGMON JIT-QueueSlotMonitor-4 lock (0x0000FFFF60002610): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "main" (J9VMThread:0x0000000000011200) 2LKREGMON JIT-QueueSlotMonitor-5 lock (0x0000FFFF600026C8): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "load-0" (J9VMThread:0x00000000003E6E00) ```
1.0
SC_Softmx_JitAot_Linux hang - https://openj9-jenkins.osuosl.org/job/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324 SC_Softmx_JitAot_Linux_1 `-Xcompressedrefs -Xjit -Xgcpolicy:gencon ` https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk17_j9_extended.system_aarch64_linux_Nightly_testList_2/324/system_test_output.tar.gz ``` STF 01:15:28.337 - **FAILED** Process jvm4 has timed out STF 01:15:28.337 - Collecting dumps for: jvm4 ``` There are core/javacore/etc. diagnostic files. The javacore files show this state: ``` 2LKREGMON JIT-QueueSlotMonitor-1 lock (0x0000FFFF94003300): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "load-1" (J9VMThread:0x000000000057C200) 3LKWAITNOTIFY "load-2" (J9VMThread:0x000000000057E700) 3LKWAITNOTIFY "load-3" (J9VMThread:0x0000000000580B00) 3LKWAITNOTIFY "load-4" (J9VMThread:0x0000000000583000) 3LKWAITNOTIFY "load-5" (J9VMThread:0x0000000000585500) 3LKWAITNOTIFY "load-6" (J9VMThread:0x0000000000587A00) 2LKREGMON JIT-QueueSlotMonitor-4 lock (0x0000FFFF60002610): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "main" (J9VMThread:0x0000000000011200) 2LKREGMON JIT-QueueSlotMonitor-5 lock (0x0000FFFF600026C8): <unowned> 3LKNOTIFYQ Waiting to be notified: 3LKWAITNOTIFY "load-0" (J9VMThread:0x00000000003E6E00) ```
test
sc softmx jitaot linux hang sc softmx jitaot linux xcompressedrefs xjit xgcpolicy gencon stf failed process has timed out stf collecting dumps for there are core javacore etc diagnostic files the javacore files show this state jit queueslotmonitor lock waiting to be notified load load load load load load jit queueslotmonitor lock waiting to be notified main jit queueslotmonitor lock waiting to be notified load
1
63,429
6,846,934,407
IssuesEvent
2017-11-13 13:54:41
Shadowss/TravianZ
https://api.github.com/repos/Shadowss/TravianZ
closed
Warning with language display
bug needs testing
``` Warning: include_once(GameEngine/lang/en.php): failed to open stream: No such file or directory in /home/u997437127/public_html/GameEngine/Technology.php on line 33 Warning: include_once(): Failed opening 'GameEngine/lang/en.php' for inclusion (include_path='.:/opt/php-7.0/pear') in /home/u997437127/public_html/GameEngine/Technology.php on line 33 ``` Need to replace `include_once($autoprefix."GameEngine/lang/en.php");` Should be `include_once($autoprefix."GameEngine/Lang/en.php");` https://github.com/Shadowss/TravianZ/blob/994cef151cf9da8b8df8d7f7f58aa60b1dc2e143/GameEngine/Technology.php#L33
1.0
Warning with language display - ``` Warning: include_once(GameEngine/lang/en.php): failed to open stream: No such file or directory in /home/u997437127/public_html/GameEngine/Technology.php on line 33 Warning: include_once(): Failed opening 'GameEngine/lang/en.php' for inclusion (include_path='.:/opt/php-7.0/pear') in /home/u997437127/public_html/GameEngine/Technology.php on line 33 ``` Need to replace `include_once($autoprefix."GameEngine/lang/en.php");` Should be `include_once($autoprefix."GameEngine/Lang/en.php");` https://github.com/Shadowss/TravianZ/blob/994cef151cf9da8b8df8d7f7f58aa60b1dc2e143/GameEngine/Technology.php#L33
test
warning with language display warning include once gameengine lang en php failed to open stream no such file or directory in home public html gameengine technology php on line warning include once failed opening gameengine lang en php for inclusion include path opt php pear in home public html gameengine technology php on line need to replace include once autoprefix gameengine lang en php should be include once autoprefix gameengine lang en php
1
352,563
32,076,821,348
IssuesEvent
2023-09-25 11:34:58
taskany-inc/issues
https://api.github.com/repos/taskany-inc/issues
closed
E2E: Project /
tests need info
- title, parent, description exists and correct - goal tab is active - settings tab is visible for owner and admin only - add goal button exists and opens new goal modal (currently is not — must be fixed)
1.0
E2E: Project / - - title, parent, description exists and correct - goal tab is active - settings tab is visible for owner and admin only - add goal button exists and opens new goal modal (currently is not — must be fixed)
test
project title parent description exists and correct goal tab is active settings tab is visible for owner and admin only add goal button exists and opens new goal modal currently is not — must be fixed
1
266,961
20,173,694,233
IssuesEvent
2022-02-10 12:44:43
halkyonio/tap
https://api.github.com/repos/halkyonio/tap
opened
Ad dthe steps to bind the quarkus app to a service
documentation enhancement
## TODO Add the steps to bind the Quarkus app to a PostgreSQL DB ## Instrcutions From Ioannis - Determine which is the resource used by TAP to install MySQL or Postgress. - Add the kubernetes-service-binding extension to your project. - Add the following config to your project: ``` quarkus.kubernetes-service-binding.services.postgresql.api-version=<api group of the postgress resource> quarkus.kubernetes-service-binding.services.postgresql.kind=<kind of the postgres resource> quarkus.kubernetes-service-binding.services.postgresql.name=<name of the postgres resource> ``` - Remove the datasource configuration you have in your application.properies (Just keep the db-kind and optionally name of the datasource) - Run mvn clean install -Dquarkus.kubernetes.deploy=true The exact steps I am using are here: https://github.com/iocanel/blog/tree/master/using-quarkus-with-the-service-binding-operator
1.0
Ad dthe steps to bind the quarkus app to a service - ## TODO Add the steps to bind the Quarkus app to a PostgreSQL DB ## Instrcutions From Ioannis - Determine which is the resource used by TAP to install MySQL or Postgress. - Add the kubernetes-service-binding extension to your project. - Add the following config to your project: ``` quarkus.kubernetes-service-binding.services.postgresql.api-version=<api group of the postgress resource> quarkus.kubernetes-service-binding.services.postgresql.kind=<kind of the postgres resource> quarkus.kubernetes-service-binding.services.postgresql.name=<name of the postgres resource> ``` - Remove the datasource configuration you have in your application.properies (Just keep the db-kind and optionally name of the datasource) - Run mvn clean install -Dquarkus.kubernetes.deploy=true The exact steps I am using are here: https://github.com/iocanel/blog/tree/master/using-quarkus-with-the-service-binding-operator
non_test
ad dthe steps to bind the quarkus app to a service todo add the steps to bind the quarkus app to a postgresql db instrcutions from ioannis determine which is the resource used by tap to install mysql or postgress add the kubernetes service binding extension to your project add the following config to your project quarkus kubernetes service binding services postgresql api version quarkus kubernetes service binding services postgresql kind quarkus kubernetes service binding services postgresql name remove the datasource configuration you have in your application properies just keep the db kind and optionally name of the datasource run mvn clean install dquarkus kubernetes deploy true the exact steps i am using are here
0
315,659
27,092,960,100
IssuesEvent
2023-02-14 22:54:20
timescale/timescaledb
https://api.github.com/repos/timescale/timescaledb
opened
[Flaky test] bgw_launcher doesn't see wait_event_changed
flaky-test
### Which test is flaky? bgw_launcher ### Since when has the test been flaky? February 14, 2023 ### Link to the failed test run https://github.com/timescale/timescaledb/actions/runs/4175600633/jobs/7233369995 ### Log output ```bash diff -u /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out --- /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out 2023-02-14 15:55:10.826147051 +0000 +++ /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out 2023-02-14 15:56:26.894945058 +0000 @@ -142,7 +142,7 @@ AND datname = :'TEST_DBNAME_2'; wait_event_changed -------------------- - t + f (1 row) -- Test stop ``` ### Reason for flakiness _No response_
1.0
[Flaky test] bgw_launcher doesn't see wait_event_changed - ### Which test is flaky? bgw_launcher ### Since when has the test been flaky? February 14, 2023 ### Link to the failed test run https://github.com/timescale/timescaledb/actions/runs/4175600633/jobs/7233369995 ### Log output ```bash diff -u /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out --- /home/runner/work/timescaledb/timescaledb/test/expected/bgw_launcher.out 2023-02-14 15:55:10.826147051 +0000 +++ /home/runner/work/timescaledb/timescaledb/build/test/results/bgw_launcher.out 2023-02-14 15:56:26.894945058 +0000 @@ -142,7 +142,7 @@ AND datname = :'TEST_DBNAME_2'; wait_event_changed -------------------- - t + f (1 row) -- Test stop ``` ### Reason for flakiness _No response_
test
bgw launcher doesn t see wait event changed which test is flaky bgw launcher since when has the test been flaky february link to the failed test run log output bash diff u home runner work timescaledb timescaledb test expected bgw launcher out home runner work timescaledb timescaledb build test results bgw launcher out home runner work timescaledb timescaledb test expected bgw launcher out home runner work timescaledb timescaledb build test results bgw launcher out and datname test dbname wait event changed t f row test stop reason for flakiness no response
1
53,466
6,330,437,025
IssuesEvent
2017-07-26 07:24:25
jiscdev/study-goal
https://api.github.com/repos/jiscdev/study-goal
closed
Photo incredibly grainy when taken from Android device
Android bug Move to testing Request Accepted
Please can we reduce the compression on the photos - too grainy at the moment. Photos selected from the library also appear rotated by 90'. Tested on Samsung S6 physical device. Photo capture: <img src="https://user-images.githubusercontent.com/5815497/28213104-8a9777d2-689c-11e7-843a-c119a528ff48.png" width="300"> Library selection: <img src="https://user-images.githubusercontent.com/5815497/28213110-8c6b48cc-689c-11e7-8d97-9bd3687941b8.png" width="300">
1.0
Photo incredibly grainy when taken from Android device - Please can we reduce the compression on the photos - too grainy at the moment. Photos selected from the library also appear rotated by 90'. Tested on Samsung S6 physical device. Photo capture: <img src="https://user-images.githubusercontent.com/5815497/28213104-8a9777d2-689c-11e7-843a-c119a528ff48.png" width="300"> Library selection: <img src="https://user-images.githubusercontent.com/5815497/28213110-8c6b48cc-689c-11e7-8d97-9bd3687941b8.png" width="300">
test
photo incredibly grainy when taken from android device please can we reduce the compression on the photos too grainy at the moment photos selected from the library also appear rotated by tested on samsung physical device photo capture library selection
1
15,813
20,713,349,761
IssuesEvent
2022-03-12 08:23:37
Creators-of-Create/Create
https://api.github.com/repos/Creators-of-Create/Create
closed
JEI potion recipes
bug needs input compatibility stale
clicking on Create potions in jei shows every recipe for automated brewing, item draining, and filling by sprout
True
JEI potion recipes - clicking on Create potions in jei shows every recipe for automated brewing, item draining, and filling by sprout
non_test
jei potion recipes clicking on create potions in jei shows every recipe for automated brewing item draining and filling by sprout
0
62,579
17,082,305,010
IssuesEvent
2021-07-08 07:24:03
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Regression in ResultQuery.fetchMap(Field, Field) with null values
T: Defect
### Expected behavior ResultQuery.fetchMap(Field, Field) should not throw NPE when value field maps to a null value. This was the behavior on 3.14.x. fetchMap's docs don't list any limitations on value field, so I assume this is unintentional. Probably also applies to other fetchMap overrides. ### Actual behavior NPE is thrown internally from HashMap.merge, at least with Java 16. java.lang.NullPointerException at java.base/java.util.HashMap.merge(HashMap.java:1359) at java.base/java.util.stream.Collectors.lambda$toMap$68(Collectors.java:1666) at java.base/java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169) at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133) at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484) at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474) at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913) at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682) at org.jooq.impl.AbstractCursor.collect(AbstractCursor.java:78) at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:358) at org.jooq.impl.ResultQueryTrait.fetchMap(ResultQueryTrait.java:850) ### Steps to reproduce the problem An example query, where TEAM.BUSINESS_ID is a nullable column. ```java var businessIds = dsl.select() .from(TEAM) .fetchMap(TEAM.ID, TEAM.BUSINESS_ID); ``` ### Versions - jOOQ: 3.15.0 - Java: 16
1.0
Regression in ResultQuery.fetchMap(Field, Field) with null values - ### Expected behavior ResultQuery.fetchMap(Field, Field) should not throw NPE when value field maps to a null value. This was the behavior on 3.14.x. fetchMap's docs don't list any limitations on value field, so I assume this is unintentional. Probably also applies to other fetchMap overrides. ### Actual behavior NPE is thrown internally from HashMap.merge, at least with Java 16. java.lang.NullPointerException at java.base/java.util.HashMap.merge(HashMap.java:1359) at java.base/java.util.stream.Collectors.lambda$toMap$68(Collectors.java:1666) at java.base/java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169) at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133) at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484) at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474) at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913) at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682) at org.jooq.impl.AbstractCursor.collect(AbstractCursor.java:78) at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:358) at org.jooq.impl.ResultQueryTrait.fetchMap(ResultQueryTrait.java:850) ### Steps to reproduce the problem An example query, where TEAM.BUSINESS_ID is a nullable column. ```java var businessIds = dsl.select() .from(TEAM) .fetchMap(TEAM.ID, TEAM.BUSINESS_ID); ``` ### Versions - jOOQ: 3.15.0 - Java: 16
non_test
regression in resultquery fetchmap field field with null values expected behavior resultquery fetchmap field field should not throw npe when value field maps to a null value this was the behavior on x fetchmap s docs don t list any limitations on value field so i assume this is unintentional probably also applies to other fetchmap overrides actual behavior npe is thrown internally from hashmap merge at least with java java lang nullpointerexception at java base java util hashmap merge hashmap java at java base java util stream collectors lambda tomap collectors java at java base java util stream reduceops accept reduceops java at java base java util iterator foreachremaining iterator java at java base java util spliterators iteratorspliterator foreachremaining spliterators java at java base java util stream abstractpipeline copyinto abstractpipeline java at java base java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java base java util stream reduceops reduceop evaluatesequential reduceops java at java base java util stream abstractpipeline evaluate abstractpipeline java at java base java util stream referencepipeline collect referencepipeline java at org jooq impl abstractcursor collect abstractcursor java at org jooq impl resultquerytrait collect resultquerytrait java at org jooq impl resultquerytrait fetchmap resultquerytrait java steps to reproduce the problem an example query where team business id is a nullable column java var businessids dsl select from team fetchmap team id team business id versions jooq java
0
118,030
9,968,483,883
IssuesEvent
2019-07-08 15:41:03
golang/go
https://api.github.com/repos/golang/go
opened
crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder
Soon Testing release-blocker
`crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder. First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself. Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6 ``` --- FAIL: TestVerifyHostnameResumed (0.10s) --- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s) tls_test.go:383: Subsequent connection unexpectedly didn't resume --- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s) tls_test.go:383: Subsequent connection unexpectedly didn't resume FAIL FAIL crypto/tls 0.828s ``` CC @agl
1.0
crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder - `crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder. First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself. Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6 ``` --- FAIL: TestVerifyHostnameResumed (0.10s) --- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s) tls_test.go:383: Subsequent connection unexpectedly didn't resume --- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s) tls_test.go:383: Subsequent connection unexpectedly didn't resume FAIL FAIL crypto/tls 0.828s ``` CC @agl
test
crypto tls testverifyhostnameresumed consistently failing in longtest builder crypto tls testverifyhostnameresumed is consistently failing in the linux longtest builder first failure was at cl which appears to be unrelated that suggests some sort of change in a non hermetic dependency or perhaps in the builder itself example fail testverifyhostnameresumed fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail fail crypto tls cc agl
1
278,131
21,058,099,640
IssuesEvent
2022-04-01 06:44:07
isabelteo/ped
https://api.github.com/repos/isabelteo/ped
opened
Standardising t/Breed
type.DocumentationBug severity.Low
![image.png](https://raw.githubusercontent.com/isabelteo/ped/main/files/515df170-29a4-4aee-8081-09a0cd34fad4.png) ![image.png](https://raw.githubusercontent.com/isabelteo/ped/main/files/c2c991a7-b4cf-4d3a-8487-8b8a430f6e56.png) Can consider to match the command summary, which uses t/Tag with t/Breed which is used above <!--session: 1648792986020-e42dd825-f9a0-404a-9fc9-346c11c52df9--> <!--Version: Web v3.4.2-->
1.0
Standardising t/Breed - ![image.png](https://raw.githubusercontent.com/isabelteo/ped/main/files/515df170-29a4-4aee-8081-09a0cd34fad4.png) ![image.png](https://raw.githubusercontent.com/isabelteo/ped/main/files/c2c991a7-b4cf-4d3a-8487-8b8a430f6e56.png) Can consider to match the command summary, which uses t/Tag with t/Breed which is used above <!--session: 1648792986020-e42dd825-f9a0-404a-9fc9-346c11c52df9--> <!--Version: Web v3.4.2-->
non_test
standardising t breed can consider to match the command summary which uses t tag with t breed which is used above
0
79,580
7,720,158,160
IssuesEvent
2018-05-23 21:54:17
NMGRL/pychron
https://api.github.com/repos/NMGRL/pychron
closed
table group
Data Reduction Enhancement Ready to close Tested OK
add a table group identifier to the tag file add set table group tag in same places as set_tag. use same set_tag code except no need to interact with database. only need to write/commit to tag file 0,0,0,1,1,1 a,b,c,d,e,f ... s=IntermediateAnalysis(d,e,f) a,b,c,s the summary age is the sum of a,b,c,s NOT a,b,c,d,e,f
1.0
table group - add a table group identifier to the tag file add set table group tag in same places as set_tag. use same set_tag code except no need to interact with database. only need to write/commit to tag file 0,0,0,1,1,1 a,b,c,d,e,f ... s=IntermediateAnalysis(d,e,f) a,b,c,s the summary age is the sum of a,b,c,s NOT a,b,c,d,e,f
test
table group add a table group identifier to the tag file add set table group tag in same places as set tag use same set tag code except no need to interact with database only need to write commit to tag file a b c d e f s intermediateanalysis d e f a b c s the summary age is the sum of a b c s not a b c d e f
1
84,053
24,209,674,760
IssuesEvent
2022-09-25 18:15:37
contour-terminal/contour
https://api.github.com/repos/contour-terminal/contour
closed
Cannot build with Qt 6.2.4
build
### Contour Terminal version master ### Installer source Github: source code cloned ### Operating System Devuan ### Architecture x86-64 ### Other Software _No response_ ### Steps to reproduce Configure with `-DCONTOUR_BUILD_WITH_QT6=ON` option. ### Expected Behavior Successfully compiling. ### Actual Behavior ``` /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp: In function ‘void BlurBehind::setEnabled(QWindow*, bool, QRegion)’: /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:37:28: error: invalid use of incomplete type ‘class QVariant’ 37 | window->setProperty("kwin_blur", region); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/qtypeinfo.h:42, from /usr/include/x86_64-linux-gnu/qt6/QtCore/qglobal.h:1431, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qtguiglobal.h:43, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:43, from /usr/include/x86_64-linux-gnu/qt6/QtGui/QWindow:1, from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.h:16, from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:14: /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:38:28: error: invalid use of incomplete type ‘class QVariant’ 38 | window->setProperty("kwin_background_region", region); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:39:57: error: cannot convert ‘int’ to ‘const QVariant&’ 39 | window->setProperty("kwin_background_contrast", 1); | ^ | | | int In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/QObject:1, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:44: /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:40:58: error: cannot convert ‘int’ to ‘const QVariant&’ 40 | window->setProperty("kwin_background_intensity", 1); | ^ | | | int /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:41:59: error: cannot convert ‘int’ to ‘const QVariant&’ 41 | window->setProperty("kwin_background_saturation", 1); | ^ | | | int /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:42:46: error: cannot convert ‘const char [38]’ to ‘const QVariant&’ 42 | window->setProperty("_MUTTER_HINTS", "blur-provider=sigma:60,brightness:0.9"); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | | | const char [38] /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:46:28: error: invalid use of incomplete type ‘const class QVariant’ 46 | window->setProperty("kwin_blur", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:47:28: error: invalid use of incomplete type ‘const class QVariant’ 47 | window->setProperty("kwin_background_region", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:48:28: error: invalid use of incomplete type ‘const class QVariant’ 48 | window->setProperty("kwin_background_contrast", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:49:28: error: invalid use of incomplete type ‘const class QVariant’ 49 | window->setProperty("kwin_background_intensity", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:50:28: error: invalid use of incomplete type ‘const class QVariant’ 50 | window->setProperty("kwin_background_saturation", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:51:28: error: invalid use of incomplete type ‘const class QVariant’ 51 | window->setProperty("_MUTTER_HINTS", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ make[2]: *** [src/contour/CMakeFiles/contour.dir/build.make:140: src/contour/CMakeFiles/contour.dir/BlurBehind.cpp.o] Error 1 make[2]: *** Waiting for unfinished jobs.... make[1]: *** [CMakeFiles/Makefile2:556: src/contour/CMakeFiles/contour.dir/all] Error 2 make: *** [Makefile:156: all] Error 2 ``` ### Additional notes _No response_
1.0
Cannot build with Qt 6.2.4 - ### Contour Terminal version master ### Installer source Github: source code cloned ### Operating System Devuan ### Architecture x86-64 ### Other Software _No response_ ### Steps to reproduce Configure with `-DCONTOUR_BUILD_WITH_QT6=ON` option. ### Expected Behavior Successfully compiling. ### Actual Behavior ``` /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp: In function ‘void BlurBehind::setEnabled(QWindow*, bool, QRegion)’: /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:37:28: error: invalid use of incomplete type ‘class QVariant’ 37 | window->setProperty("kwin_blur", region); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/qtypeinfo.h:42, from /usr/include/x86_64-linux-gnu/qt6/QtCore/qglobal.h:1431, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qtguiglobal.h:43, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:43, from /usr/include/x86_64-linux-gnu/qt6/QtGui/QWindow:1, from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.h:16, from /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:14: /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:38:28: error: invalid use of incomplete type ‘class QVariant’ 38 | window->setProperty("kwin_background_region", region); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:39:57: error: cannot convert ‘int’ to ‘const QVariant&’ 39 | window->setProperty("kwin_background_contrast", 1); | ^ | | | int In file included from /usr/include/x86_64-linux-gnu/qt6/QtCore/QObject:1, from /usr/include/x86_64-linux-gnu/qt6/QtGui/qwindow.h:44: /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:40:58: error: cannot convert ‘int’ to ‘const QVariant&’ 40 | window->setProperty("kwin_background_intensity", 1); | ^ | | | int /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:41:59: error: cannot convert ‘int’ to ‘const QVariant&’ 41 | window->setProperty("kwin_background_saturation", 1); | ^ | | | int /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:42:46: error: cannot convert ‘const char [38]’ to ‘const QVariant&’ 42 | window->setProperty("_MUTTER_HINTS", "blur-provider=sigma:60,brightness:0.9"); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | | | const char [38] /usr/include/x86_64-linux-gnu/qt6/QtCore/qobject.h:376:56: note: initializing argument 2 of ‘bool QObject::setProperty(const char*, const QVariant&)’ 376 | bool setProperty(const char *name, const QVariant &value); | ~~~~~~~~~~~~~~~~^~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:46:28: error: invalid use of incomplete type ‘const class QVariant’ 46 | window->setProperty("kwin_blur", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:47:28: error: invalid use of incomplete type ‘const class QVariant’ 47 | window->setProperty("kwin_background_region", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:48:28: error: invalid use of incomplete type ‘const class QVariant’ 48 | window->setProperty("kwin_background_contrast", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:49:28: error: invalid use of incomplete type ‘const class QVariant’ 49 | window->setProperty("kwin_background_intensity", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:50:28: error: invalid use of incomplete type ‘const class QVariant’ 50 | window->setProperty("kwin_background_saturation", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ /home/dataman/Projects/C/---Terminals/contour/src/contour/BlurBehind.cpp:51:28: error: invalid use of incomplete type ‘const class QVariant’ 51 | window->setProperty("_MUTTER_HINTS", {}); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /usr/include/x86_64-linux-gnu/qt6/QtCore/qcontainerfwd.h:72:7: note: forward declaration of ‘class QVariant’ 72 | class QVariant; | ^~~~~~~~ make[2]: *** [src/contour/CMakeFiles/contour.dir/build.make:140: src/contour/CMakeFiles/contour.dir/BlurBehind.cpp.o] Error 1 make[2]: *** Waiting for unfinished jobs.... make[1]: *** [CMakeFiles/Makefile2:556: src/contour/CMakeFiles/contour.dir/all] Error 2 make: *** [Makefile:156: all] Error 2 ``` ### Additional notes _No response_
non_test
cannot build with qt contour terminal version master installer source github source code cloned operating system devuan architecture other software no response steps to reproduce configure with dcontour build with on option expected behavior successfully compiling actual behavior home dataman projects c terminals contour src contour blurbehind cpp in function ‘void blurbehind setenabled qwindow bool qregion ’ home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘class qvariant’ window setproperty kwin blur region in file included from usr include linux gnu qtcore qtypeinfo h from usr include linux gnu qtcore qglobal h from usr include linux gnu qtgui qtguiglobal h from usr include linux gnu qtgui qwindow h from usr include linux gnu qtgui qwindow from home dataman projects c terminals contour src contour blurbehind h from home dataman projects c terminals contour src contour blurbehind cpp usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘class qvariant’ window setproperty kwin background region region usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background contrast int in file included from usr include linux gnu qtcore qobject from usr include linux gnu qtgui qwindow h usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background intensity int usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘int’ to ‘const qvariant ’ window setproperty kwin background saturation int usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error cannot convert ‘const char ’ to ‘const qvariant ’ window setproperty mutter hints blur provider sigma brightness const char usr include linux gnu qtcore qobject h note initializing argument of ‘bool qobject setproperty const char const qvariant ’ bool setproperty const char name const qvariant value home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin blur usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background region usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background contrast usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background intensity usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty kwin background saturation usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant home dataman projects c terminals contour src contour blurbehind cpp error invalid use of incomplete type ‘const class qvariant’ window setproperty mutter hints usr include linux gnu qtcore qcontainerfwd h note forward declaration of ‘class qvariant’ class qvariant make error make waiting for unfinished jobs make error make error additional notes no response
0
194,033
14,667,706,651
IssuesEvent
2020-12-29 19:19:28
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
rkand4/vuejs: pkg/http/listener_test.go; 40 LoC
fresh small test
Found a possible issue in [rkand4/vuejs](https://www.github.com/rkand4/vuejs) at [pkg/http/listener_test.go](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable i used in defer or goroutine at line 522 [Click here to see the code in its original context.](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543) <details> <summary>Click here to show the 40 line(s) of Go which triggered the analyzer.</summary> ```go for i, testCase := range testCases { listener, err := newHTTPListener( testCase.serverAddrs, testCase.tlsConfig, time.Duration(0), time.Duration(0), time.Duration(0), nil, nil, errorFunc, ) if err != nil { t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err) } go func() { serverConn, aerr := listener.Accept() if aerr == nil { t.Fatalf("Test %d: accept: expected = <error>, got = <nil>", i+1) } if serverConn != nil { t.Fatalf("Test %d: accept: server expected = <nil>, got = %v", i+1, serverConn) } }() for _, serverAddr := range testCase.serverAddrs { conn, err := net.Dial("tcp", serverAddr) if err != nil { t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err) } if _, err = io.WriteString(conn, testCase.request); err != nil { t.Fatalf("Test %d: request send: expected = <nil>, got = %v", i+1, err) } conn.Close() } listener.Close() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: e790bf122760e516384839f20608d7b79b1d344a
1.0
rkand4/vuejs: pkg/http/listener_test.go; 40 LoC - Found a possible issue in [rkand4/vuejs](https://www.github.com/rkand4/vuejs) at [pkg/http/listener_test.go](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable i used in defer or goroutine at line 522 [Click here to see the code in its original context.](https://github.com/rkand4/vuejs/blob/e790bf122760e516384839f20608d7b79b1d344a/pkg/http/listener_test.go#L504-L543) <details> <summary>Click here to show the 40 line(s) of Go which triggered the analyzer.</summary> ```go for i, testCase := range testCases { listener, err := newHTTPListener( testCase.serverAddrs, testCase.tlsConfig, time.Duration(0), time.Duration(0), time.Duration(0), nil, nil, errorFunc, ) if err != nil { t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err) } go func() { serverConn, aerr := listener.Accept() if aerr == nil { t.Fatalf("Test %d: accept: expected = <error>, got = <nil>", i+1) } if serverConn != nil { t.Fatalf("Test %d: accept: server expected = <nil>, got = %v", i+1, serverConn) } }() for _, serverAddr := range testCase.serverAddrs { conn, err := net.Dial("tcp", serverAddr) if err != nil { t.Fatalf("Test %d: error: expected = <nil>, got = %v", i+1, err) } if _, err = io.WriteString(conn, testCase.request); err != nil { t.Fatalf("Test %d: request send: expected = <nil>, got = %v", i+1, err) } conn.Close() } listener.Close() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: e790bf122760e516384839f20608d7b79b1d344a
test
vuejs pkg http listener test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable i used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for i testcase range testcases listener err newhttplistener testcase serveraddrs testcase tlsconfig time duration time duration time duration nil nil errorfunc if err nil t fatalf test d error expected got v i err go func serverconn aerr listener accept if aerr nil t fatalf test d accept expected got i if serverconn nil t fatalf test d accept server expected got v i serverconn for serveraddr range testcase serveraddrs conn err net dial tcp serveraddr if err nil t fatalf test d error expected got v i err if err io writestring conn testcase request err nil t fatalf test d request send expected got v i err conn close listener close leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
44,518
18,069,191,714
IssuesEvent
2021-09-20 23:25:07
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Search] Use new elasticsearch-js client in data plugin search service
Feature:Search enhancement loe:hours Team:AppServices impact:low
Related: https://github.com/elastic/kibana/issues/35508 Now that https://github.com/elastic/kibana/pull/69905 is merged, we have access to the new elasticsearch-js client (which updates Typescript support for ES as well as a handful of other benefits), we should update the data plugin search service to use the new client rather than the legacy client. This would open up the ability for other enhancements (see https://github.com/elastic/kibana/issues/70675).
1.0
[Search] Use new elasticsearch-js client in data plugin search service - Related: https://github.com/elastic/kibana/issues/35508 Now that https://github.com/elastic/kibana/pull/69905 is merged, we have access to the new elasticsearch-js client (which updates Typescript support for ES as well as a handful of other benefits), we should update the data plugin search service to use the new client rather than the legacy client. This would open up the ability for other enhancements (see https://github.com/elastic/kibana/issues/70675).
non_test
use new elasticsearch js client in data plugin search service related now that is merged we have access to the new elasticsearch js client which updates typescript support for es as well as a handful of other benefits we should update the data plugin search service to use the new client rather than the legacy client this would open up the ability for other enhancements see
0
221,907
17,376,543,935
IssuesEvent
2021-07-30 22:36:17
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
No error info prompts when switching proxy setting from 'Do not use proxy' to 'Use app proxy settings' without configuring the correct proxy
:beetle: regression :gear: proxy :heavy_check_mark: merged 🧪 testing
**Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210723.3 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: From running test cases **Regression From**: Previous release (1.20.1) ## Steps to Reproduce ## 1. Launch Storage Explorer -> Make sure the proxy setting is 'Do not use proxy' -> Close Storage Explorer. 2. Download Fiddler. 3. Configure Fiddler from https://docs.microsoft.com/en-us/azure/storage/common/storage-explorer-troubleshooting?tabs=Windows%2C2004#network-logs. 4. Open Storage Explorer -> Switch to 'Use app proxy settings'. 5. Check whether an error info bar prompts. ## Expected Experience ## An error info bar prompts. ![image](https://user-images.githubusercontent.com/41351993/126762651-f4c8fe78-630a-482d-a96a-713e7805fc13.png) ## Actual Experience ## No error info bar prompts. ## Additional Context ## There is no traffic for operations like 'upload, delete blobs...' after configuring the correct settings for 'Use app proxy settings'.
1.0
No error info prompts when switching proxy setting from 'Do not use proxy' to 'Use app proxy settings' without configuring the correct proxy - **Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210723.3 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: From running test cases **Regression From**: Previous release (1.20.1) ## Steps to Reproduce ## 1. Launch Storage Explorer -> Make sure the proxy setting is 'Do not use proxy' -> Close Storage Explorer. 2. Download Fiddler. 3. Configure Fiddler from https://docs.microsoft.com/en-us/azure/storage/common/storage-explorer-troubleshooting?tabs=Windows%2C2004#network-logs. 4. Open Storage Explorer -> Switch to 'Use app proxy settings'. 5. Check whether an error info bar prompts. ## Expected Experience ## An error info bar prompts. ![image](https://user-images.githubusercontent.com/41351993/126762651-f4c8fe78-630a-482d-a96a-713e7805fc13.png) ## Actual Experience ## No error info bar prompts. ## Additional Context ## There is no traffic for operations like 'upload, delete blobs...' after configuring the correct settings for 'Use app proxy settings'.
test
no error info prompts when switching proxy setting from do not use proxy to use app proxy settings without configuring the correct proxy storage explorer version dev build number branch main platform os windows architecture how found from running test cases regression from previous release steps to reproduce launch storage explorer make sure the proxy setting is do not use proxy close storage explorer download fiddler configure fiddler from open storage explorer switch to use app proxy settings check whether an error info bar prompts expected experience an error info bar prompts actual experience no error info bar prompts additional context there is no traffic for operations like upload delete blobs after configuring the correct settings for use app proxy settings
1
103,531
16,602,910,551
IssuesEvent
2021-06-01 22:15:11
gms-ws-sandbox/nibrs
https://api.github.com/repos/gms-ws-sandbox/nibrs
opened
CVE-2018-1272 (High) detected in spring-core-4.3.11.RELEASE.jar
security vulnerability
## CVE-2018-1272 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-4.3.11.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.11.RELEASE/spring-core-4.3.11.RELEASE.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/spring-core-4.3.11.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-core-4.3.11.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles. <p>Publish Date: 2018-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272>CVE-2018-1272</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2018-1272">https://tanzu.vmware.com/security/cve-2018-1272</a></p> <p>Release Date: 2018-04-06</p> <p>Fix Resolution: org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"4.3.11.RELEASE","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:4.3.11.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1272 (High) detected in spring-core-4.3.11.RELEASE.jar - ## CVE-2018-1272 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-4.3.11.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-core/4.3.11.RELEASE/spring-core-4.3.11.RELEASE.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/spring-core-4.3.11.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-core-4.3.11.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles. <p>Publish Date: 2018-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272>CVE-2018-1272</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2018-1272">https://tanzu.vmware.com/security/cve-2018-1272</a></p> <p>Release Date: 2018-04-06</p> <p>Fix Resolution: org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"4.3.11.RELEASE","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:4.3.11.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-core:4.3.15.RELEASE,5.0.5.RELEASE;org.springframework:spring-web:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1272","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in spring core release jar cve high severity vulnerability vulnerable library spring core release jar spring core library home page a href path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library home wss scanner repository org springframework spring core release spring core release jar nibrs tools nibrs fbi service target nibrs fbi service web inf lib spring core release jar dependency hierarchy x spring core release jar vulnerable library found in head commit a href found in base branch master vulnerability details spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring core release release org springframework spring web release release isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org springframework spring core release isminimumfixversionavailable true minimumfixversion org springframework spring core release release org springframework spring web release release basebranches vulnerabilityidentifier cve vulnerabilitydetails spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles vulnerabilityurl
0
188,921
14,479,842,662
IssuesEvent
2020-12-10 10:21:02
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
触发构建时,支持填入构建信息
area/ci/api area/ci/backend area/ci/frontend kind/feat/product priority/critical-urgent stage/test stage/uat test/passed
### 期望 手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。 如果没有填入,平台给构建信息置入默认值:“手动触发” 后续将作为能替代构建号的重要信息作为构建索引。 ![image](https://user-images.githubusercontent.com/10875468/97959945-60979900-1deb-11eb-9270-ec3fc9e60c7c.png) ### 注意 需考虑以下场景: 代码库触发 - 用最新commit msg置入构建信息; 远程触发 - `curl`指令内需包含`BK_CI_BUILD_MSG`。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值; 定时触发 - 默认“定时触发”; 子流水线触发 - 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入${BK_CI_BUILD_MSG},则替换默认值; 服务触发 - 默认“服务触发” 关联需求:#2850
2.0
触发构建时,支持填入构建信息 - ### 期望 手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。 如果没有填入,平台给构建信息置入默认值:“手动触发” 后续将作为能替代构建号的重要信息作为构建索引。 ![image](https://user-images.githubusercontent.com/10875468/97959945-60979900-1deb-11eb-9270-ec3fc9e60c7c.png) ### 注意 需考虑以下场景: 代码库触发 - 用最新commit msg置入构建信息; 远程触发 - `curl`指令内需包含`BK_CI_BUILD_MSG`。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值; 定时触发 - 默认“定时触发”; 子流水线触发 - 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入${BK_CI_BUILD_MSG},则替换默认值; 服务触发 - 默认“服务触发” 关联需求:#2850
test
触发构建时,支持填入构建信息 期望 手动触发流水线时,需填入该次构建的信息,用以简单描述该次构建的目的、备注等。 如果没有填入,平台给构建信息置入默认值:“手动触发” 后续将作为能替代构建号的重要信息作为构建索引。 注意 需考虑以下场景: 代码库触发 用最新commit msg置入构建信息; 远程触发 curl 指令内需包含 bk ci build msg 。若值为空,将默认值“远程触发”置入构建信息;若值不为空,则替换默认值; 定时触发 默认“定时触发”; 子流水线触发 默认“流水线调用触发”,若用户在【子流水线调用】插件内传入 bk ci build msg ,则替换默认值; 服务触发 默认“服务触发” 关联需求:
1
117,437
9,935,111,014
IssuesEvent
2019-07-02 15:48:05
somaliaims/Somali-AIMSUI
https://api.github.com/repos/somaliaims/Somali-AIMSUI
closed
sector mapping on input add sector
july testing
The goal of the AIMS is to store all projects with a Somali sector - all the other sector types are just there to assist users entering data that comes from another sector type. So when the sector is imported from IATI, or another AIMS project, we automatically select the sector type that it is coming from. When it is a manual entry - we allow the user to select the sector type they want to enter. In both cases, the UI then needs to show what default (in this case Somali) sector the input sector type will be stored as (and allow the user to adjust if they want to). If there is no mapping, then the user should be asked to select a Somali sector. The user then confirms the Somali sector that gets stored.
1.0
sector mapping on input add sector - The goal of the AIMS is to store all projects with a Somali sector - all the other sector types are just there to assist users entering data that comes from another sector type. So when the sector is imported from IATI, or another AIMS project, we automatically select the sector type that it is coming from. When it is a manual entry - we allow the user to select the sector type they want to enter. In both cases, the UI then needs to show what default (in this case Somali) sector the input sector type will be stored as (and allow the user to adjust if they want to). If there is no mapping, then the user should be asked to select a Somali sector. The user then confirms the Somali sector that gets stored.
test
sector mapping on input add sector the goal of the aims is to store all projects with a somali sector all the other sector types are just there to assist users entering data that comes from another sector type so when the sector is imported from iati or another aims project we automatically select the sector type that it is coming from when it is a manual entry we allow the user to select the sector type they want to enter in both cases the ui then needs to show what default in this case somali sector the input sector type will be stored as and allow the user to adjust if they want to if there is no mapping then the user should be asked to select a somali sector the user then confirms the somali sector that gets stored
1
104,954
9,013,179,094
IssuesEvent
2019-02-05 18:48:30
NillerMedDild/Enigmatica2Expert
https://api.github.com/repos/NillerMedDild/Enigmatica2Expert
closed
Fluids not Despawning properly after source block removed
needs testing
I've noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal.....in particular the pool water from the slime islands and the healing spa water from the alps biomes. ## Thank you for helping us improve. ##### _Please follow the issue template, and do not delete any part of it._ _Failure to comply will result in the issue being closed._ **Issue Description** **Modpack version (Do not use "latest")** **Logs** If applicable, please upload your crash report, or latest.log to [gist](https://gist.github.com/). **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
1.0
Fluids not Despawning properly after source block removed - I've noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal.....in particular the pool water from the slime islands and the healing spa water from the alps biomes. ## Thank you for helping us improve. ##### _Please follow the issue template, and do not delete any part of it._ _Failure to comply will result in the issue being closed._ **Issue Description** **Modpack version (Do not use "latest")** **Logs** If applicable, please upload your crash report, or latest.log to [gist](https://gist.github.com/). **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
test
fluids not despawning properly after source block removed i ve noticed that flowing fluids other than water and lave that have their source block removed or blocked are taking way longer to despawn than normal in particular the pool water from the slime islands and the healing spa water from the alps biomes thank you for helping us improve please follow the issue template and do not delete any part of it failure to comply will result in the issue being closed issue description modpack version do not use latest logs if applicable please upload your crash report or latest log to screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here
1
285,399
24,663,808,692
IssuesEvent
2022-10-18 08:47:36
rancher/cis-operator
https://api.github.com/repos/rancher/cis-operator
closed
ClusterRoles for cis-operator and scan job pods are too permissive
[zube]: To Test team/area3 team/infracloud feature/charts-cis-benchmark team/rke2
## Issue The existing ClusteRoles ([here](https://github.com/rancher/charts/blob/dev-v2.6/charts/rancher-cis-benchmark/2.1.0/templates/rbac.yaml)) attached to the ServiceAccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported. For instance: ``` groups: - id: exploitable-controle text: "" checks: - id: exploitable-check text: "" audit: > kubectl delete ns kube-system ``` ## Potential Solution In the Helm chart, the roles attached to the SAs should be scope down to specific required functions. We've been testing and these roles seem to work pretty well for all existing scans + custom scans: #### Operator ClusterRole and Role ```yaml apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: name: cis-operator-clusterrole rules: - apiGroups: - "cis.cattle.io" resources: - "*" verbs: - "*" - apiGroups: - "" resources: - "pods" - "services" - "configmaps" - "nodes" verbs: - "list" - "create" - "update" - "watch" - apiGroups: - "batch" resources: - "jobs" verbs: - "list" - "create" - "update" - "watch" --- apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: name: cis-operator-role namespace: cis-operator-system rules: - apiGroups: - "" resources: - "services" verbs: - "watch" - "list" - "get" - "patch" - apiGroups: - "batch" resources: - "jobs" verbs: - "watch" - "list" - "get" - "delete" - apiGroups: - "" resources: - "configmaps" - "pods" verbs: - "watch" - "list" - "get" - "delete" ``` #### Scan ClusterRole & Role ```yaml apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: name: cis-scan-clusterrole rules: - apiGroups: - "*" resources: - "*" verbs: - "list" - "describe" - "get" --- apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: name: cis-scan-role namespace: cis-operator-system rules: - apiGroups: - "" resources: - "pods" verbs: - "patch" - "get" - "list" - "describe" - apiGroups: - "apps" resources: - "daemonsets" verbs: - "*" - apiGroups: - "" resources: - "secrets" - "configmaps" verbs: - "update" - "create" ```
1.0
ClusterRoles for cis-operator and scan job pods are too permissive - ## Issue The existing ClusteRoles ([here](https://github.com/rancher/charts/blob/dev-v2.6/charts/rancher-cis-benchmark/2.1.0/templates/rbac.yaml)) attached to the ServiceAccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported. For instance: ``` groups: - id: exploitable-controle text: "" checks: - id: exploitable-check text: "" audit: > kubectl delete ns kube-system ``` ## Potential Solution In the Helm chart, the roles attached to the SAs should be scope down to specific required functions. We've been testing and these roles seem to work pretty well for all existing scans + custom scans: #### Operator ClusterRole and Role ```yaml apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: name: cis-operator-clusterrole rules: - apiGroups: - "cis.cattle.io" resources: - "*" verbs: - "*" - apiGroups: - "" resources: - "pods" - "services" - "configmaps" - "nodes" verbs: - "list" - "create" - "update" - "watch" - apiGroups: - "batch" resources: - "jobs" verbs: - "list" - "create" - "update" - "watch" --- apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: name: cis-operator-role namespace: cis-operator-system rules: - apiGroups: - "" resources: - "services" verbs: - "watch" - "list" - "get" - "patch" - apiGroups: - "batch" resources: - "jobs" verbs: - "watch" - "list" - "get" - "delete" - apiGroups: - "" resources: - "configmaps" - "pods" verbs: - "watch" - "list" - "get" - "delete" ``` #### Scan ClusterRole & Role ```yaml apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: name: cis-scan-clusterrole rules: - apiGroups: - "*" resources: - "*" verbs: - "list" - "describe" - "get" --- apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: name: cis-scan-role namespace: cis-operator-system rules: - apiGroups: - "" resources: - "pods" verbs: - "patch" - "get" - "list" - "describe" - apiGroups: - "apps" resources: - "daemonsets" verbs: - "*" - apiGroups: - "" resources: - "secrets" - "configmaps" verbs: - "update" - "create" ```
test
clusterroles for cis operator and scan job pods are too permissive issue the existing clusteroles attached to the serviceaccounts are too permissive and can be used to destroy a cluster if a custom profile with exploitable audit commands is imported for instance groups id exploitable controle text checks id exploitable check text audit kubectl delete ns kube system potential solution in the helm chart the roles attached to the sas should be scope down to specific required functions we ve been testing and these roles seem to work pretty well for all existing scans custom scans operator clusterrole and role yaml apiversion rbac authorization io kind clusterrole metadata name cis operator clusterrole rules apigroups cis cattle io resources verbs apigroups resources pods services configmaps nodes verbs list create update watch apigroups batch resources jobs verbs list create update watch apiversion rbac authorization io kind role metadata name cis operator role namespace cis operator system rules apigroups resources services verbs watch list get patch apigroups batch resources jobs verbs watch list get delete apigroups resources configmaps pods verbs watch list get delete scan clusterrole role yaml apiversion rbac authorization io kind clusterrole metadata name cis scan clusterrole rules apigroups resources verbs list describe get apiversion rbac authorization io kind role metadata name cis scan role namespace cis operator system rules apigroups resources pods verbs patch get list describe apigroups apps resources daemonsets verbs apigroups resources secrets configmaps verbs update create
1
140,179
5,398,323,310
IssuesEvent
2017-02-27 16:40:51
metacpan/metacpan-web
https://api.github.com/repos/metacpan/metacpan-web
closed
Allow users to view their user_id
priority:Low Volunteer needed
Please let the site's users view their user id, so that they will then be able to ask 3rd-party sites & services to fetch their MetaCPAN favorites or interact in other ways. One such service is https://www.perlmodules.net/, where retrieving one's own list of favorites would be quite useful. So, exposing the user id under https://metacpan.org/account/profile might be the most appropriate place.
1.0
Allow users to view their user_id - Please let the site's users view their user id, so that they will then be able to ask 3rd-party sites & services to fetch their MetaCPAN favorites or interact in other ways. One such service is https://www.perlmodules.net/, where retrieving one's own list of favorites would be quite useful. So, exposing the user id under https://metacpan.org/account/profile might be the most appropriate place.
non_test
allow users to view their user id please let the site s users view their user id so that they will then be able to ask party sites services to fetch their metacpan favorites or interact in other ways one such service is where retrieving one s own list of favorites would be quite useful so exposing the user id under might be the most appropriate place
0
57,888
6,559,861,391
IssuesEvent
2017-09-07 06:53:43
nodejs/node
https://api.github.com/repos/nodejs/node
closed
OSX: intermittent failure: async-hooks/test-callback-error
async_hooks macos test
* **Version**: master * **Platform**: OSX * **Subsystem**: async hooks <!-- Enter your issue details below this comment. --> Failure in run for unrelated change: https://ci.nodejs.org/job/node-test-commit-osx/12183/nodes=osx1010/console ``` not ok 76 async-hooks/test-callback-error --- duration_ms: 15.551 severity: fail stack: |- start case 1 end case 1: 114.761ms start case 2 end case 2: 115.167ms start case 3 end case 3: 8.221ms Error: test_callback_abort at ActivityCollector.initHooks.oninit.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:36:45) at ActivityCollector.oninit (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:509:15) at ActivityCollector._init (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/init-hooks.js:182:10) at emitInitNative (async_hooks.js:446:43) at Object.emitInitScript [as emitInit] (async_hooks.js:349:3) at Object.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:38:17) at Module._compile (module.js:549:30) at Object.Module._extensions..js (module.js:560:10) at Module.load (module.js:483:32) at tryModuleLoad (module.js:446:12) 1: node::Abort() [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 6: 0x31be80046fd ```
1.0
OSX: intermittent failure: async-hooks/test-callback-error - * **Version**: master * **Platform**: OSX * **Subsystem**: async hooks <!-- Enter your issue details below this comment. --> Failure in run for unrelated change: https://ci.nodejs.org/job/node-test-commit-osx/12183/nodes=osx1010/console ``` not ok 76 async-hooks/test-callback-error --- duration_ms: 15.551 severity: fail stack: |- start case 1 end case 1: 114.761ms start case 2 end case 2: 115.167ms start case 3 end case 3: 8.221ms Error: test_callback_abort at ActivityCollector.initHooks.oninit.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:36:45) at ActivityCollector.oninit (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:509:15) at ActivityCollector._init (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/init-hooks.js:182:10) at emitInitNative (async_hooks.js:446:43) at Object.emitInitScript [as emitInit] (async_hooks.js:349:3) at Object.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/async-hooks/test-callback-error.js:38:17) at Module._compile (module.js:549:30) at Object.Module._extensions..js (module.js:560:10) at Module.load (module.js:483:32) at tryModuleLoad (module.js:446:12) 1: node::Abort() [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/out/Release/node] 6: 0x31be80046fd ```
test
osx intermittent failure async hooks test callback error version master platform osx subsystem async hooks failure in run for unrelated change not ok async hooks test callback error duration ms severity fail stack start case end case start case end case start case end case error test callback abort at activitycollector inithooks oninit common mustcall users iojs build workspace node test commit osx nodes test async hooks test callback error js at activitycollector oninit users iojs build workspace node test commit osx nodes test common index js at activitycollector init users iojs build workspace node test commit osx nodes test async hooks init hooks js at emitinitnative async hooks js at object emitinitscript async hooks js at object users iojs build workspace node test commit osx nodes test async hooks test callback error js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js node abort node chdir functioncallbackinfo const internal functioncallbackarguments call void functioncallbackinfo const internal maybehandle internal anonymous namespace handleapicallhelper internal isolate internal handle internal handle internal handle internal handle internal builtinarguments internal builtin impl handleapicall internal builtinarguments internal isolate
1
309,052
26,649,126,881
IssuesEvent
2023-01-25 12:22:17
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql/tests: TestRandomSyntaxGeneration failed
C-test-failure O-robot branch-master T-sql-sessions
sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=artifacts#/) on master @ [642afd6e8c8d1f967da4aa0c3c08e2bdc0495100](https://github.com/cockroachdb/cockroach/commits/642afd6e8c8d1f967da4aa0c3c08e2bdc0495100): ``` crypto/tls.(*atLeastReader).Read(0xc05661a8b8, {0xc04a583d40?, 0x0?, 0x389488168619053b?}) GOROOT/src/crypto/tls/conn.go:787 +0x3d fp=0xc0339d0938 sp=0xc0339d08f0 pc=0x86799d bytes.(*Buffer).ReadFrom(0xc0569a3078, {0x6c2e660, 0xc05661a8b8}) GOROOT/src/bytes/buffer.go:202 +0x98 fp=0xc0339d0990 sp=0xc0339d0938 pc=0x565858 crypto/tls.(*Conn).readFromUntil(0xc0569a2e00, {0x6c3dd20?, 0xc04bd20b18}, 0x80deb1fe72be5d74?) GOROOT/src/crypto/tls/conn.go:809 +0xe5 fp=0xc0339d09d0 sp=0xc0339d0990 pc=0x867b85 crypto/tls.(*Conn).readRecordOrCCS(0xc0569a2e00, 0x0) GOROOT/src/crypto/tls/conn.go:616 +0x116 fp=0xc0339d0d58 sp=0xc0339d09d0 pc=0x864fd6 crypto/tls.(*Conn).readRecord(...) GOROOT/src/crypto/tls/conn.go:582 crypto/tls.(*Conn).readHandshake(0xc0569a2e00) GOROOT/src/crypto/tls/conn.go:1019 +0x6d fp=0xc0339d0e38 sp=0xc0339d0d58 pc=0x868ded crypto/tls.(*serverHandshakeStateTLS13).readClientCertificate(0xc0339d11c0) GOROOT/src/crypto/tls/handshake_server_tls13.go:787 +0x65 fp=0xc0339d10c0 sp=0xc0339d0e38 pc=0x8948e5 crypto/tls.(*serverHandshakeStateTLS13).handshake(0xc0339d11c0) GOROOT/src/crypto/tls/handshake_server_tls13.go:78 +0xb6 fp=0xc0339d10e0 sp=0xc0339d10c0 pc=0x890bb6 crypto/tls.(*Conn).serverHandshake(0xc0569a2e00, {0x6c75730, 0xc0569a0b00}) GOROOT/src/crypto/tls/handshake_server.go:54 +0xd0 fp=0xc0339d12b8 sp=0xc0339d10e0 pc=0x88b550 crypto/tls.(*Conn).serverHandshake-fm({0x6c75730?, 0xc0569a0b00?}) <autogenerated>:1 +0x39 fp=0xc0339d12e0 sp=0xc0339d12b8 pc=0x89f4d9 crypto/tls.(*Conn).handshakeContext(0xc0569a2e00, {0x6c75768, 0xc000080048}) GOROOT/src/crypto/tls/conn.go:1462 +0x32f fp=0xc0339d14a8 sp=0xc0339d12e0 pc=0x86b6cf crypto/tls.(*Conn).HandshakeContext(...) GOROOT/src/crypto/tls/conn.go:1405 crypto/tls.(*Conn).Handshake(...) GOROOT/src/crypto/tls/conn.go:1389 crypto/tls.(*Conn).Read(0xc0569a2e00, {0xc0569c4438, 0x4, 0x493f10?}) GOROOT/src/crypto/tls/conn.go:1274 +0x65 fp=0xc0339d1518 sp=0xc0339d14a8 pc=0x86ab25 io.ReadAtLeast({0x7eff005e27f8, 0xc0569a2e00}, {0xc0569c4438, 0x4, 0x4}, 0x4) GOROOT/src/io/io.go:332 +0x9a fp=0xc0339d1560 sp=0xc0339d1518 pc=0x50ccba io.ReadFull(...) GOROOT/src/io/io.go:351 github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase.(*ReadBuffer).ReadUntypedMsg(0xc0569c4420, {0x7eff005e27f8, 0xc0569a2e00}) github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase/encoding.go:155 +0x65 fp=0xc0339d1720 sp=0xc0339d1560 pc=0x27e6c85 github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).readVersion(0xc000cf23c0, {0x7eff005e27f8, 0xc0569a2e00}) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1371 +0x127 fp=0xc0339d17a8 sp=0xc0339d1720 pc=0x38ce567 github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).maybeUpgradeToSecureConn(0xc000cf23c0, {0x0?, 0xffc?}, {0x6cbb9a8, 0xc04bd20b18}, 0x2, 0x6c3dd20?, 0xc0339d1bc0) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1315 +0x35d fp=0xc0339d1878 sp=0xc0339d17a8 pc=0x38cdd9d github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).ServeConn(0xc000cf23c0, {0x6c75730, 0xc056b06880}, {0x6cbb9a8?, 0xc04bd20b18}, 0x6a?) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:811 +0x59e fp=0xc0339d1e20 sp=0xc0339d1878 pc=0x38ca83e github.com/cockroachdb/cockroach/pkg/server.(*SQLServer).startServeSQL.func1.1({0x6c757d8?, 0xc0568207b0?}, {0x6cbb9a8, 0xc04bd20b18}) github.com/cockroachdb/cockroach/pkg/server/server_sql.go:1568 +0xc8 fp=0xc0339d1eb0 sp=0xc0339d1e20 pc=0x3b95648 github.com/cockroachdb/cockroach/pkg/util/netutil.(*TCPServer).ServeWith.func1({0x6c757d8, 0xc0568207b0}) github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:188 +0xc7 fp=0xc0339d1f30 sp=0xc0339d1eb0 pc=0x17d96a7 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2() github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 +0x146 fp=0xc0339d1fe0 sp=0xc0339d1f30 pc=0x10fdae6 runtime.goexit() GOROOT/src/runtime/asm_amd64.s:1594 +0x1 fp=0xc0339d1fe8 sp=0xc0339d1fe0 pc=0x4c4021 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:461 +0x43b ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #89363 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.2.0] - #87572 sql/tests: TestRandomSyntaxGeneration failed [DROP OWNED BY timeout] [C-test-failure O-robot T-sql-schema branch-release-22.2] - #77893 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.1] - #74271 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.2] - #65210 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.1] </p> </details> /cc @cockroachdb/sql-sessions <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-23617
1.0
sql/tests: TestRandomSyntaxGeneration failed - sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8208463?buildTab=artifacts#/) on master @ [642afd6e8c8d1f967da4aa0c3c08e2bdc0495100](https://github.com/cockroachdb/cockroach/commits/642afd6e8c8d1f967da4aa0c3c08e2bdc0495100): ``` crypto/tls.(*atLeastReader).Read(0xc05661a8b8, {0xc04a583d40?, 0x0?, 0x389488168619053b?}) GOROOT/src/crypto/tls/conn.go:787 +0x3d fp=0xc0339d0938 sp=0xc0339d08f0 pc=0x86799d bytes.(*Buffer).ReadFrom(0xc0569a3078, {0x6c2e660, 0xc05661a8b8}) GOROOT/src/bytes/buffer.go:202 +0x98 fp=0xc0339d0990 sp=0xc0339d0938 pc=0x565858 crypto/tls.(*Conn).readFromUntil(0xc0569a2e00, {0x6c3dd20?, 0xc04bd20b18}, 0x80deb1fe72be5d74?) GOROOT/src/crypto/tls/conn.go:809 +0xe5 fp=0xc0339d09d0 sp=0xc0339d0990 pc=0x867b85 crypto/tls.(*Conn).readRecordOrCCS(0xc0569a2e00, 0x0) GOROOT/src/crypto/tls/conn.go:616 +0x116 fp=0xc0339d0d58 sp=0xc0339d09d0 pc=0x864fd6 crypto/tls.(*Conn).readRecord(...) GOROOT/src/crypto/tls/conn.go:582 crypto/tls.(*Conn).readHandshake(0xc0569a2e00) GOROOT/src/crypto/tls/conn.go:1019 +0x6d fp=0xc0339d0e38 sp=0xc0339d0d58 pc=0x868ded crypto/tls.(*serverHandshakeStateTLS13).readClientCertificate(0xc0339d11c0) GOROOT/src/crypto/tls/handshake_server_tls13.go:787 +0x65 fp=0xc0339d10c0 sp=0xc0339d0e38 pc=0x8948e5 crypto/tls.(*serverHandshakeStateTLS13).handshake(0xc0339d11c0) GOROOT/src/crypto/tls/handshake_server_tls13.go:78 +0xb6 fp=0xc0339d10e0 sp=0xc0339d10c0 pc=0x890bb6 crypto/tls.(*Conn).serverHandshake(0xc0569a2e00, {0x6c75730, 0xc0569a0b00}) GOROOT/src/crypto/tls/handshake_server.go:54 +0xd0 fp=0xc0339d12b8 sp=0xc0339d10e0 pc=0x88b550 crypto/tls.(*Conn).serverHandshake-fm({0x6c75730?, 0xc0569a0b00?}) <autogenerated>:1 +0x39 fp=0xc0339d12e0 sp=0xc0339d12b8 pc=0x89f4d9 crypto/tls.(*Conn).handshakeContext(0xc0569a2e00, {0x6c75768, 0xc000080048}) GOROOT/src/crypto/tls/conn.go:1462 +0x32f fp=0xc0339d14a8 sp=0xc0339d12e0 pc=0x86b6cf crypto/tls.(*Conn).HandshakeContext(...) GOROOT/src/crypto/tls/conn.go:1405 crypto/tls.(*Conn).Handshake(...) GOROOT/src/crypto/tls/conn.go:1389 crypto/tls.(*Conn).Read(0xc0569a2e00, {0xc0569c4438, 0x4, 0x493f10?}) GOROOT/src/crypto/tls/conn.go:1274 +0x65 fp=0xc0339d1518 sp=0xc0339d14a8 pc=0x86ab25 io.ReadAtLeast({0x7eff005e27f8, 0xc0569a2e00}, {0xc0569c4438, 0x4, 0x4}, 0x4) GOROOT/src/io/io.go:332 +0x9a fp=0xc0339d1560 sp=0xc0339d1518 pc=0x50ccba io.ReadFull(...) GOROOT/src/io/io.go:351 github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase.(*ReadBuffer).ReadUntypedMsg(0xc0569c4420, {0x7eff005e27f8, 0xc0569a2e00}) github.com/cockroachdb/cockroach/pkg/sql/pgwire/pgwirebase/encoding.go:155 +0x65 fp=0xc0339d1720 sp=0xc0339d1560 pc=0x27e6c85 github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).readVersion(0xc000cf23c0, {0x7eff005e27f8, 0xc0569a2e00}) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1371 +0x127 fp=0xc0339d17a8 sp=0xc0339d1720 pc=0x38ce567 github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).maybeUpgradeToSecureConn(0xc000cf23c0, {0x0?, 0xffc?}, {0x6cbb9a8, 0xc04bd20b18}, 0x2, 0x6c3dd20?, 0xc0339d1bc0) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:1315 +0x35d fp=0xc0339d1878 sp=0xc0339d17a8 pc=0x38cdd9d github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*Server).ServeConn(0xc000cf23c0, {0x6c75730, 0xc056b06880}, {0x6cbb9a8?, 0xc04bd20b18}, 0x6a?) github.com/cockroachdb/cockroach/pkg/sql/pgwire/server.go:811 +0x59e fp=0xc0339d1e20 sp=0xc0339d1878 pc=0x38ca83e github.com/cockroachdb/cockroach/pkg/server.(*SQLServer).startServeSQL.func1.1({0x6c757d8?, 0xc0568207b0?}, {0x6cbb9a8, 0xc04bd20b18}) github.com/cockroachdb/cockroach/pkg/server/server_sql.go:1568 +0xc8 fp=0xc0339d1eb0 sp=0xc0339d1e20 pc=0x3b95648 github.com/cockroachdb/cockroach/pkg/util/netutil.(*TCPServer).ServeWith.func1({0x6c757d8, 0xc0568207b0}) github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:188 +0xc7 fp=0xc0339d1f30 sp=0xc0339d1eb0 pc=0x17d96a7 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2() github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 +0x146 fp=0xc0339d1fe0 sp=0xc0339d1f30 pc=0x10fdae6 runtime.goexit() GOROOT/src/runtime/asm_amd64.s:1594 +0x1 fp=0xc0339d1fe8 sp=0xc0339d1fe0 pc=0x4c4021 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:461 +0x43b ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #89363 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.2.0] - #87572 sql/tests: TestRandomSyntaxGeneration failed [DROP OWNED BY timeout] [C-test-failure O-robot T-sql-schema branch-release-22.2] - #77893 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot T-sql-sessions branch-release-22.1] - #74271 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.2] - #65210 sql/tests: TestRandomSyntaxGeneration failed [C-test-failure O-robot branch-release-21.1] </p> </details> /cc @cockroachdb/sql-sessions <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-23617
test
sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration with on master crypto tls atleastreader read goroot src crypto tls conn go fp sp pc bytes buffer readfrom goroot src bytes buffer go fp sp pc crypto tls conn readfromuntil goroot src crypto tls conn go fp sp pc crypto tls conn readrecordorccs goroot src crypto tls conn go fp sp pc crypto tls conn readrecord goroot src crypto tls conn go crypto tls conn readhandshake goroot src crypto tls conn go fp sp pc crypto tls readclientcertificate goroot src crypto tls handshake server go fp sp pc crypto tls handshake goroot src crypto tls handshake server go fp sp pc crypto tls conn serverhandshake goroot src crypto tls handshake server go fp sp pc crypto tls conn serverhandshake fm fp sp pc crypto tls conn handshakecontext goroot src crypto tls conn go fp sp pc crypto tls conn handshakecontext goroot src crypto tls conn go crypto tls conn handshake goroot src crypto tls conn go crypto tls conn read goroot src crypto tls conn go fp sp pc io readatleast goroot src io io go fp sp pc io readfull goroot src io io go github com cockroachdb cockroach pkg sql pgwire pgwirebase readbuffer readuntypedmsg github com cockroachdb cockroach pkg sql pgwire pgwirebase encoding go fp sp pc github com cockroachdb cockroach pkg sql pgwire server readversion github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg sql pgwire server maybeupgradetosecureconn github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg sql pgwire server serveconn github com cockroachdb cockroach pkg sql pgwire server go fp sp pc github com cockroachdb cockroach pkg server sqlserver startservesql github com cockroachdb cockroach pkg server server sql go fp sp pc github com cockroachdb cockroach pkg util netutil tcpserver servewith github com cockroachdb cockroach pkg util netutil net go fp sp pc github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go fp sp pc runtime goexit goroot src runtime asm s fp sp pc created by github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go help see also same failure on other branches sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration failed cc cockroachdb sql sessions jira issue crdb
1
43,974
5,579,233,275
IssuesEvent
2017-03-28 14:13:14
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
Test: new commands to increase/decrease view size
testplan-item
Refs: https://github.com/Microsoft/vscode/issues/22645 - [x] anyOS - @weinand (macOS) - [x] anyOS - @egamma New commands (`workbench.action.increaseViewSize` and `workbench.action.decreaseViewSize`) allow to increase or decrease a view (sidebar, editor or panel) size with keyboard only. Verify * the commands work as advertised on the view that has current keyboard focus * you cannot break the layout (e.g. minimal and maximal dimensions are respected)
1.0
Test: new commands to increase/decrease view size - Refs: https://github.com/Microsoft/vscode/issues/22645 - [x] anyOS - @weinand (macOS) - [x] anyOS - @egamma New commands (`workbench.action.increaseViewSize` and `workbench.action.decreaseViewSize`) allow to increase or decrease a view (sidebar, editor or panel) size with keyboard only. Verify * the commands work as advertised on the view that has current keyboard focus * you cannot break the layout (e.g. minimal and maximal dimensions are respected)
test
test new commands to increase decrease view size refs anyos weinand macos anyos egamma new commands workbench action increaseviewsize and workbench action decreaseviewsize allow to increase or decrease a view sidebar editor or panel size with keyboard only verify the commands work as advertised on the view that has current keyboard focus you cannot break the layout e g minimal and maximal dimensions are respected
1
516,307
14,979,063,594
IssuesEvent
2021-01-28 11:42:40
Uninett/Argus-frontend
https://api.github.com/repos/Uninett/Argus-frontend
opened
Bug: login background isn't reset to paper after login when using userpass-auth
bug polish priority: low size: small usability
**Describe the bug** The login view sets the background using the useBackground() hook, but never reset's it. When the user logs using FEIDE this isn't a problem because FEIDE authentication causes a page reload. When the user logs in using userpass-auth however, it will NOT cause a reload, and the background stays cyan. **To Reproduce** Steps to reproduce the behavior: 1. Go to /login 2. Enter username and password 3. Press enter 4. Observe that the background used in the login view is still set **Expected behavior** The background should be reset to the default for material ui, which is paper **Screenshots** ![2021-01-28-124018_1243x938_scrot](https://user-images.githubusercontent.com/1580755/106133787-0302b680-6166-11eb-8ccf-ac8f0e820d2c.png) **Solution** Use the useEffect hook's cleanup property to unset the background. This could probably be done in the useBackground hook.
1.0
Bug: login background isn't reset to paper after login when using userpass-auth - **Describe the bug** The login view sets the background using the useBackground() hook, but never reset's it. When the user logs using FEIDE this isn't a problem because FEIDE authentication causes a page reload. When the user logs in using userpass-auth however, it will NOT cause a reload, and the background stays cyan. **To Reproduce** Steps to reproduce the behavior: 1. Go to /login 2. Enter username and password 3. Press enter 4. Observe that the background used in the login view is still set **Expected behavior** The background should be reset to the default for material ui, which is paper **Screenshots** ![2021-01-28-124018_1243x938_scrot](https://user-images.githubusercontent.com/1580755/106133787-0302b680-6166-11eb-8ccf-ac8f0e820d2c.png) **Solution** Use the useEffect hook's cleanup property to unset the background. This could probably be done in the useBackground hook.
non_test
bug login background isn t reset to paper after login when using userpass auth describe the bug the login view sets the background using the usebackground hook but never reset s it when the user logs using feide this isn t a problem because feide authentication causes a page reload when the user logs in using userpass auth however it will not cause a reload and the background stays cyan to reproduce steps to reproduce the behavior go to login enter username and password press enter observe that the background used in the login view is still set expected behavior the background should be reset to the default for material ui which is paper screenshots solution use the useeffect hook s cleanup property to unset the background this could probably be done in the usebackground hook
0
331,908
29,170,456,303
IssuesEvent
2023-05-19 00:58:12
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_build_tuple_unpack_dynamic_shapes_static_default (__main__.StaticDefaultDynamicShapesMiscTests)
triaged module: flaky-tests skipped module: dynamo
Platforms: asan, linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack_dynamic_shapes_static_default&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_build_tuple_unpack_dynamic_shapes_static_default` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py`
1.0
DISABLED test_build_tuple_unpack_dynamic_shapes_static_default (__main__.StaticDefaultDynamicShapesMiscTests) - Platforms: asan, linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack_dynamic_shapes_static_default&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_build_tuple_unpack_dynamic_shapes_static_default` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py`
test
disabled test build tuple unpack dynamic shapes static default main staticdefaultdynamicshapesmisctests platforms asan linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test build tuple unpack dynamic shapes static default there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path dynamo test dynamic shapes py or dynamo test dynamic shapes py
1
116,960
9,904,496,122
IssuesEvent
2019-06-27 09:18:29
HarmanU/test-utility-app
https://api.github.com/repos/HarmanU/test-utility-app
opened
Test Plan PDF generator
Test Plan Creator enhancement
Exports a PDF for the test plan can use similar method to the Jira card exporter to added customisation
1.0
Test Plan PDF generator - Exports a PDF for the test plan can use similar method to the Jira card exporter to added customisation
test
test plan pdf generator exports a pdf for the test plan can use similar method to the jira card exporter to added customisation
1
256,081
22,036,520,069
IssuesEvent
2022-05-28 17:17:14
Cookie-AutoDelete/Cookie-AutoDelete
https://api.github.com/repos/Cookie-AutoDelete/Cookie-AutoDelete
opened
[Bug] LocalStorage is cleaned even when "Keep LocalStorage" is enabled for a site
untested bug/issue
### Acknowledgements - [X] I acknowledge that I have read the above items ### Describe the bug [Assuming I've understood things correctly](https://github.com/Cookie-AutoDelete/Cookie-AutoDelete/issues/982#issuecomment-1140297632), there issue I'm having is this: With LocalStorage enabled in the Settings under the _Other Browsing Data Cleanup_ options and also _Keep LocalStorage_ ticked for a particular site's whitelist options, every time I start the browser, the LocalStorage is cleaned. At least, this is all I can assume because even though LocalStorage is _not_ mentioned in the logs after restarting the browser, and even with everything ticked for the site's options, the only way to keep myself signed in is to _disable_ LocalStorage in the global settings. The site is https://mail.proton.me and the rule in use is `*.proton.me` (which should cover account.proton.me and other related/necessary sites to maintain being logged-in). All the expression's "Keep" options are enabled. ### To Reproduce 1. Enable LocalStorage Cleanup in the global settings. 2. Add `*.proton.me` expression, tick every "Keep" option. 3. Log-in to the service, wait till it loads, close the browser. 4. Re-open the browser and be asked to sign in again. ### Expected Behavior Step 4 should not ask you to sign in again, i.e. keep the LocalStorage when _Keep LocalStorage_ is ticked for the expression. ### Screenshots _No response_ ### System Info - Operating System (OS) Windows 11 ### System Info - Browser Info Brave 1.39.111 Chromium: 102.0.5005.61 (Official Build) (64-bit) ### System Info - CookieAutoDelete Version 3.7.0 ### Additional Context _No response_
1.0
[Bug] LocalStorage is cleaned even when "Keep LocalStorage" is enabled for a site - ### Acknowledgements - [X] I acknowledge that I have read the above items ### Describe the bug [Assuming I've understood things correctly](https://github.com/Cookie-AutoDelete/Cookie-AutoDelete/issues/982#issuecomment-1140297632), there issue I'm having is this: With LocalStorage enabled in the Settings under the _Other Browsing Data Cleanup_ options and also _Keep LocalStorage_ ticked for a particular site's whitelist options, every time I start the browser, the LocalStorage is cleaned. At least, this is all I can assume because even though LocalStorage is _not_ mentioned in the logs after restarting the browser, and even with everything ticked for the site's options, the only way to keep myself signed in is to _disable_ LocalStorage in the global settings. The site is https://mail.proton.me and the rule in use is `*.proton.me` (which should cover account.proton.me and other related/necessary sites to maintain being logged-in). All the expression's "Keep" options are enabled. ### To Reproduce 1. Enable LocalStorage Cleanup in the global settings. 2. Add `*.proton.me` expression, tick every "Keep" option. 3. Log-in to the service, wait till it loads, close the browser. 4. Re-open the browser and be asked to sign in again. ### Expected Behavior Step 4 should not ask you to sign in again, i.e. keep the LocalStorage when _Keep LocalStorage_ is ticked for the expression. ### Screenshots _No response_ ### System Info - Operating System (OS) Windows 11 ### System Info - Browser Info Brave 1.39.111 Chromium: 102.0.5005.61 (Official Build) (64-bit) ### System Info - CookieAutoDelete Version 3.7.0 ### Additional Context _No response_
test
localstorage is cleaned even when keep localstorage is enabled for a site acknowledgements i acknowledge that i have read the above items describe the bug there issue i m having is this with localstorage enabled in the settings under the other browsing data cleanup options and also keep localstorage ticked for a particular site s whitelist options every time i start the browser the localstorage is cleaned at least this is all i can assume because even though localstorage is not mentioned in the logs after restarting the browser and even with everything ticked for the site s options the only way to keep myself signed in is to disable localstorage in the global settings the site is and the rule in use is proton me which should cover account proton me and other related necessary sites to maintain being logged in all the expression s keep options are enabled to reproduce enable localstorage cleanup in the global settings add proton me expression tick every keep option log in to the service wait till it loads close the browser re open the browser and be asked to sign in again expected behavior step should not ask you to sign in again i e keep the localstorage when keep localstorage is ticked for the expression screenshots no response system info operating system os windows system info browser info brave chromium official build bit system info cookieautodelete version additional context no response
1
29,776
4,535,896,269
IssuesEvent
2016-09-08 18:44:15
appium/appium
https://api.github.com/repos/appium/appium
opened
/wd/hub/session/:sessionId/appium/device/lock not implemented
iOS XCUITest
`/wd/hub/session/:sessionId/appium/device/lock` not implemented in xcuitest driver
1.0
/wd/hub/session/:sessionId/appium/device/lock not implemented - `/wd/hub/session/:sessionId/appium/device/lock` not implemented in xcuitest driver
test
wd hub session sessionid appium device lock not implemented wd hub session sessionid appium device lock not implemented in xcuitest driver
1
766,064
26,870,308,981
IssuesEvent
2023-02-04 11:30:11
AUBGTheHUB/spa-website-2022
https://api.github.com/repos/AUBGTheHUB/spa-website-2022
closed
Footer
high priority api frontend SPA HackAUBG
## Brief description: Design and create the Footer component ## User Story Should have instagram, facebook, email, address? . Should be reusable in HackAUBG and other pages -> a prop should be passed which would either render OG design (used in main page and Jobs page) or Hack design (difference is in the background color). you can probably set a property with a default value (e.g. `usedInHackPage = false`). Contact me @asynchroza after developing the initial footer and we will decide together how to approach this issue.
1.0
Footer - ## Brief description: Design and create the Footer component ## User Story Should have instagram, facebook, email, address? . Should be reusable in HackAUBG and other pages -> a prop should be passed which would either render OG design (used in main page and Jobs page) or Hack design (difference is in the background color). you can probably set a property with a default value (e.g. `usedInHackPage = false`). Contact me @asynchroza after developing the initial footer and we will decide together how to approach this issue.
non_test
footer brief description design and create the footer component user story should have instagram facebook email address should be reusable in hackaubg and other pages a prop should be passed which would either render og design used in main page and jobs page or hack design difference is in the background color you can probably set a property with a default value e g usedinhackpage false contact me asynchroza after developing the initial footer and we will decide together how to approach this issue
0
303,392
26,204,302,721
IssuesEvent
2023-01-03 20:46:27
BoBAdministration/QA-Bug-Reports
https://api.github.com/repos/BoBAdministration/QA-Bug-Reports
closed
Emote wheel default chats display two "Yes" options
can't reproduce Fixed-PendingTesting
**Describe the Bug** In the wheel, you should have a yes and no option, instead, both options both say "yes". **To Reproduce** Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones. 1. Log in to any server 2. Spawn as anything 3. Use the emote wheel to get to the chat buttons 4. Both options will say yes, or both options will say yes & no. However, both options will say yes. **Expected behavior** There would be a yes and no option. **Actual behavior** Both say yes. **Screenshots & Video** ![BeastsOfBermuda Development 12_29_2022 8_30_00 PM](https://user-images.githubusercontent.com/48891088/210025809-e8fe1e67-5e47-4819-9f4f-de7c33cc8dc0.png) https://user-images.githubusercontent.com/48891088/210025796-e48eaf52-775f-4021-be2f-725fde43910c.mp4 **Branch Version** Live & Tester **Character Information** Paras 1.2 **Additional Information** Titania
1.0
Emote wheel default chats display two "Yes" options - **Describe the Bug** In the wheel, you should have a yes and no option, instead, both options both say "yes". **To Reproduce** Steps to reproduce the behavior in detail. Please include ALL steps, even menial ones. 1. Log in to any server 2. Spawn as anything 3. Use the emote wheel to get to the chat buttons 4. Both options will say yes, or both options will say yes & no. However, both options will say yes. **Expected behavior** There would be a yes and no option. **Actual behavior** Both say yes. **Screenshots & Video** ![BeastsOfBermuda Development 12_29_2022 8_30_00 PM](https://user-images.githubusercontent.com/48891088/210025809-e8fe1e67-5e47-4819-9f4f-de7c33cc8dc0.png) https://user-images.githubusercontent.com/48891088/210025796-e48eaf52-775f-4021-be2f-725fde43910c.mp4 **Branch Version** Live & Tester **Character Information** Paras 1.2 **Additional Information** Titania
test
emote wheel default chats display two yes options describe the bug in the wheel you should have a yes and no option instead both options both say yes to reproduce steps to reproduce the behavior in detail please include all steps even menial ones log in to any server spawn as anything use the emote wheel to get to the chat buttons both options will say yes or both options will say yes no however both options will say yes expected behavior there would be a yes and no option actual behavior both say yes screenshots video branch version live tester character information paras additional information titania
1
108,024
9,256,946,129
IssuesEvent
2019-03-17 00:06:06
rstoneback/pysat
https://api.github.com/repos/rstoneback/pysat
opened
Minimize code duplication in testing instruments
testing
There are several areas in the testing instruments that can be combined to reduce duplication of code. - `testaddx` instruments can be folded into the main pysat_testing.py instrument, using tags to change data output. - `testsmall` instruments can be potentially be folded back into pysat_testing.py, as `tag='100'` would create default behaviour. - With current parameters, the test constellations `test_diff`, `test_diff2`, and `test_diff_small` should perform the exact same function. @rstoneback, is this a placeholder for future tests, or can these be combined?
1.0
Minimize code duplication in testing instruments - There are several areas in the testing instruments that can be combined to reduce duplication of code. - `testaddx` instruments can be folded into the main pysat_testing.py instrument, using tags to change data output. - `testsmall` instruments can be potentially be folded back into pysat_testing.py, as `tag='100'` would create default behaviour. - With current parameters, the test constellations `test_diff`, `test_diff2`, and `test_diff_small` should perform the exact same function. @rstoneback, is this a placeholder for future tests, or can these be combined?
test
minimize code duplication in testing instruments there are several areas in the testing instruments that can be combined to reduce duplication of code testaddx instruments can be folded into the main pysat testing py instrument using tags to change data output testsmall instruments can be potentially be folded back into pysat testing py as tag would create default behaviour with current parameters the test constellations test diff test and test diff small should perform the exact same function rstoneback is this a placeholder for future tests or can these be combined
1
79,187
15,586,110,645
IssuesEvent
2021-03-18 01:11:49
mibo32/fitbit-api-example-java
https://api.github.com/repos/mibo32/fitbit-api-example-java
opened
CVE-2020-24750 (High) detected in jackson-databind-2.8.1.jar
security vulnerability
## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: fitbit-api-example-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-28</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-24750 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: fitbit-api-example-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-28</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file fitbit api example java pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
148,547
11,856,388,624
IssuesEvent
2020-03-25 07:22:55
onaio/reveal-frontend
https://api.github.com/repos/onaio/reveal-frontend
opened
Test blocks with jest spyon lead to test failure if they appear above other tests.
tests
In `/src/containers/pages/FocusInvestigation/map/active/tests/index.test.tsx` move the last test block (the one with @todo comment) to the top and run the tests again. They will fail
1.0
Test blocks with jest spyon lead to test failure if they appear above other tests. - In `/src/containers/pages/FocusInvestigation/map/active/tests/index.test.tsx` move the last test block (the one with @todo comment) to the top and run the tests again. They will fail
test
test blocks with jest spyon lead to test failure if they appear above other tests in src containers pages focusinvestigation map active tests index test tsx move the last test block the one with todo comment to the top and run the tests again they will fail
1
279,461
24,227,110,307
IssuesEvent
2022-09-26 15:10:28
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] XPackRestIT test {p0=transform/transforms_unattended/Test unattended put and start wildcard} failing
>test-failure :ml/Transform Team:ML
**Build scan:** https://gradle-enterprise.elastic.co/s/mc6mzbewge5rq/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_unattended%2FTest%20unattended%20put%20and%20start%20wildcard%7D **Reproduction line:** `./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_unattended/Test unattended put and start wildcard}" -Dtests.seed=EF18978E741FE47A -Dtests.locale=no -Dtests.timezone=Antarctica/Vostok -Druntime.java=17` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_unattended/Test%20unattended%20put%20and%20start%20wildcard%7D **Failure excerpt:** ``` java.lang.AssertionError: Failure at [transform/transforms_unattended:14]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}] at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484) at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}] at org.junit.Assert.fail(Assert.java:88) at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484) at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
1.0
[CI] XPackRestIT test {p0=transform/transforms_unattended/Test unattended put and start wildcard} failing - **Build scan:** https://gradle-enterprise.elastic.co/s/mc6mzbewge5rq/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_unattended%2FTest%20unattended%20put%20and%20start%20wildcard%7D **Reproduction line:** `./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_unattended/Test unattended put and start wildcard}" -Dtests.seed=EF18978E741FE47A -Dtests.locale=no -Dtests.timezone=Antarctica/Vostok -Druntime.java=17` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_unattended/Test%20unattended%20put%20and%20start%20wildcard%7D **Failure excerpt:** ``` java.lang.AssertionError: Failure at [transform/transforms_unattended:14]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}] at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484) at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [transform-unattended] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:330)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:306)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1362)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1447)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1427)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onFailure(TransportTasksAction.java:114)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$3(TransportStopTransformAction.java:247)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onFailure(ActionListener.java:170)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:164)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.6.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:825)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}] at org.junit.Assert.fail(Assert.java:88) at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499) at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:484) at jdk.internal.reflect.GeneratedMethodAccessor13.invoke(null:-1) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
test
xpackrestit test transform transforms unattended test unattended put and start wildcard failing build scan reproduction line gradlew x pack plugin yamlresttest tests org elasticsearch xpack test rest xpackrestit test transform transforms unattended test unattended put and start wildcard dtests seed dtests locale no dtests timezone antarctica vostok druntime java applicable branches main reproduces locally no failure history failure excerpt java lang assertionerror failure at expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect invoke null at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java caused by java lang assertionerror expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onfailure transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onfailure actionlistener java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at org junit assert fail assert java at org elasticsearch test rest yaml section dosection execute dosection java at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect invoke null at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
1
301,732
26,092,644,629
IssuesEvent
2022-12-26 13:53:17
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Jest Tests.x-pack/plugins/cases/public/components/all_cases - AllCasesListGeneric Actions Row actions should disable row actions when selecting a case
failed-test
A test failed on a tracked branch ``` TestingLibraryElementError: Unable to find an element by: [data-test-subj="checkboxSelectRow-basic-case-id"] Ignored nodes: comments, script, style <body class="" > <div /> </body> at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19 at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1010:31) at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12) at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14) at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1009:12) at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28) at new Promise (<anonymous>) at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10) at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40) at runMicrotasks (<anonymous>) at runNextTicks (node:internal/process/task_queues:61:5) at listOnTimeout (node:internal/timers:528:9) at processTimers (node:internal/timers:502:7) at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3) at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21) at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19) at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16) at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25182#01854e8e-ad88-4aeb-95ef-5e047b0791fa) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/all_cases","test.name":"AllCasesListGeneric Actions Row actions should disable row actions when selecting a case","test.failCount":1}} -->
1.0
Failing test: Jest Tests.x-pack/plugins/cases/public/components/all_cases - AllCasesListGeneric Actions Row actions should disable row actions when selecting a case - A test failed on a tracked branch ``` TestingLibraryElementError: Unable to find an element by: [data-test-subj="checkboxSelectRow-basic-case-id"] Ignored nodes: comments, script, style <body class="" > <div /> </body> at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19 at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1010:31) at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12) at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14) at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/all_cases/all_cases_list.test.tsx:1009:12) at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28) at new Promise (<anonymous>) at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10) at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40) at runMicrotasks (<anonymous>) at runNextTicks (node:internal/process/task_queues:61:5) at listOnTimeout (node:internal/timers:528:9) at processTimers (node:internal/timers:502:7) at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3) at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21) at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19) at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16) at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-1154c761d1f6f4cb/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25182#01854e8e-ad88-4aeb-95ef-5e047b0791fa) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/all_cases","test.name":"AllCasesListGeneric Actions Row actions should disable row actions when selecting a case","test.failCount":1}} -->
test
failing test jest tests x pack plugins cases public components all cases allcaseslistgeneric actions row actions should disable row actions when selecting a case a test failed on a tracked branch testinglibraryelementerror unable to find an element by ignored nodes comments script style body class at object getelementerror var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist config js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at getbytestid var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components all cases all cases list test tsx at batchedupdates var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom development js at act var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom test utils development js at object var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components all cases all cases list test tsx at promise then completed var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at new promise at callasynccircusfn var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at callcircustest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runmicrotasks at runnextticks node internal process task queues at listontimeout node internal timers at processtimers node internal timers at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at run var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runandtransformresultstojestformat var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js first failure
1
451,490
13,036,859,590
IssuesEvent
2020-07-28 12:57:51
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1636] Web elections: graphs don't match
Category: Elections Website Category: Web Priority: High Status: Fixed
1. This is how graph looks in the preview for comments ![bandicam 2020-07-02 17-41-20-690](https://user-images.githubusercontent.com/27898520/86360909-7ace3880-bc8c-11ea-9677-e699a15b21fe.jpg) See, this one has bars for actions 2. After saving we don't have this bar here no more ![bandicam 2020-07-02 17-41-30-184](https://user-images.githubusercontent.com/27898520/86361523-6b032400-bc8d-11ea-851f-5742e8fead7e.jpg)
1.0
[0.9.0 staging-1636] Web elections: graphs don't match - 1. This is how graph looks in the preview for comments ![bandicam 2020-07-02 17-41-20-690](https://user-images.githubusercontent.com/27898520/86360909-7ace3880-bc8c-11ea-9677-e699a15b21fe.jpg) See, this one has bars for actions 2. After saving we don't have this bar here no more ![bandicam 2020-07-02 17-41-30-184](https://user-images.githubusercontent.com/27898520/86361523-6b032400-bc8d-11ea-851f-5742e8fead7e.jpg)
non_test
web elections graphs don t match this is how graph looks in the preview for comments see this one has bars for actions after saving we don t have this bar here no more
0
342,873
30,642,309,636
IssuesEvent
2023-07-24 23:35:09
microsoft/vscode-remote-release
https://api.github.com/repos/microsoft/vscode-remote-release
opened
Test: local port forwarding using tunnels
testplan-item
Refs https://github.com/microsoft/vscode-internalbacklog/issues/4382 - [ ] macOS - [ ] windows Complexity: 3 Authors: @connor4312, @alexr00 --- This iteration we built the first version of "local forwarding", which allows users to forward their local ports and make them accessible to the internet using Dev Tunnels. 1. Install the latest prerelease of the Remote Tunnels extension (note: this is [still in PR](https://github.com/microsoft/vscode-remote-tunnels/pull/661) at time of writing, so may not be published by Zurich's Tuesday morning) 2. Set `"remote.tunnels.experimental.forwarding": true` in your user settings, and reload the window 3. The "Ports" view should now appear, even though you are not connected to a remote. 4. You should be able to forward a port and go to the resulting URL to view anything you have running on that port locally. 5. Test out forwarding multiple ports from multiple VS Code windows. 6. Capture the Output from the "Remote Tunnels" extension if you have any issues
1.0
Test: local port forwarding using tunnels - Refs https://github.com/microsoft/vscode-internalbacklog/issues/4382 - [ ] macOS - [ ] windows Complexity: 3 Authors: @connor4312, @alexr00 --- This iteration we built the first version of "local forwarding", which allows users to forward their local ports and make them accessible to the internet using Dev Tunnels. 1. Install the latest prerelease of the Remote Tunnels extension (note: this is [still in PR](https://github.com/microsoft/vscode-remote-tunnels/pull/661) at time of writing, so may not be published by Zurich's Tuesday morning) 2. Set `"remote.tunnels.experimental.forwarding": true` in your user settings, and reload the window 3. The "Ports" view should now appear, even though you are not connected to a remote. 4. You should be able to forward a port and go to the resulting URL to view anything you have running on that port locally. 5. Test out forwarding multiple ports from multiple VS Code windows. 6. Capture the Output from the "Remote Tunnels" extension if you have any issues
test
test local port forwarding using tunnels refs macos windows complexity authors this iteration we built the first version of local forwarding which allows users to forward their local ports and make them accessible to the internet using dev tunnels install the latest prerelease of the remote tunnels extension note this is at time of writing so may not be published by zurich s tuesday morning set remote tunnels experimental forwarding true in your user settings and reload the window the ports view should now appear even though you are not connected to a remote you should be able to forward a port and go to the resulting url to view anything you have running on that port locally test out forwarding multiple ports from multiple vs code windows capture the output from the remote tunnels extension if you have any issues
1
4,238
20,999,654,346
IssuesEvent
2022-03-29 16:13:25
jxk20/nlb_goodreads_searcher
https://api.github.com/repos/jxk20/nlb_goodreads_searcher
closed
Set up github CI/CD pipeline
maintainability
- [x] Do tests for `client` - [x] Check for code coverage - [ ] Do tests for `server`
True
Set up github CI/CD pipeline - - [x] Do tests for `client` - [x] Check for code coverage - [ ] Do tests for `server`
non_test
set up github ci cd pipeline do tests for client check for code coverage do tests for server
0
319,204
27,355,908,428
IssuesEvent
2023-02-27 12:51:52
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
lorax-templates v4.2.1 (r4.2)
r4.2-host-cur-test
Update of lorax-templates to v4.2.1 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-lorax-templates/commit/969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 [Changes since previous version](https://github.com/QubesOS/qubes-lorax-templates/compare/v4.2.0...v4.2.1): QubesOS/qubes-lorax-templates@969b6dc version 4.2.1 QubesOS/qubes-lorax-templates@5e0324f Handle kernel versions with .qubes.fc suffix, instead of .fc.qubes QubesOS/qubes-lorax-templates@df7326a Use inst.repo= instead of inst.stage2= QubesOS/qubes-lorax-templates@19dd3fe Unbreak optional sshd QubesOS/qubes-lorax-templates@81c0646 Install artwork package QubesOS/qubes-lorax-templates@4debf10 Clarify kernel-latest boot entry Referenced issues: QubesOS/qubes-issues#6792 If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current all` (available 5 days from now) * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 security-testing` You can choose subset of distributions like: * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
1.0
lorax-templates v4.2.1 (r4.2) - Update of lorax-templates to v4.2.1 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-lorax-templates/commit/969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 [Changes since previous version](https://github.com/QubesOS/qubes-lorax-templates/compare/v4.2.0...v4.2.1): QubesOS/qubes-lorax-templates@969b6dc version 4.2.1 QubesOS/qubes-lorax-templates@5e0324f Handle kernel versions with .qubes.fc suffix, instead of .fc.qubes QubesOS/qubes-lorax-templates@df7326a Use inst.repo= instead of inst.stage2= QubesOS/qubes-lorax-templates@19dd3fe Unbreak optional sshd QubesOS/qubes-lorax-templates@81c0646 Install artwork package QubesOS/qubes-lorax-templates@4debf10 Clarify kernel-latest boot entry Referenced issues: QubesOS/qubes-issues#6792 If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current all` (available 5 days from now) * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 security-testing` You can choose subset of distributions like: * `Upload-component r4.2 lorax-templates 969b6dc0def1ed5b1f7a025ac3ec0461c7eb1799 current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
test
lorax templates update of lorax templates to for qubes see comments below for details and build status from commit qubesos qubes lorax templates version qubesos qubes lorax templates handle kernel versions with qubes fc suffix instead of fc qubes qubesos qubes lorax templates use inst repo instead of inst qubesos qubes lorax templates unbreak optional sshd qubesos qubes lorax templates install artwork package qubesos qubes lorax templates clarify kernel latest boot entry referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload component lorax templates current all available days from now upload component lorax templates security testing you can choose subset of distributions like upload component lorax templates current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
1
127,779
10,482,223,270
IssuesEvent
2019-09-24 11:31:57
chainer/chainer
https://api.github.com/repos/chainer/chainer
closed
flaky test: `chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction::test_double_backward`
cat:test pr-ongoing prio:high
Possibly related to #7946 https://jenkins.preferred.jp/job/chainer/job/cupy_pr/347/TEST=chainer-py35,label=mn1-p100/console >`FAIL tests/chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always_param_25_{dilate=1, nobias=False, x_dtype=float16, contiguous=None, W_dtype=float16, groups=2, cover_all=False}::test_double_backward` ``` 18:48:25 E chainer.testing.function_link.FunctionTestError: Parameterized test failed. 18:48:25 E 18:48:25 E Base test method: TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always.test_double_backward 18:48:25 E Test parameters: 18:48:25 E dilate: 1 18:48:25 E nobias: False 18:48:25 E x_dtype: <class 'numpy.float16'> 18:48:25 E contiguous: None 18:48:25 E W_dtype: <class 'numpy.float16'> 18:48:25 E groups: 2 18:48:25 E cover_all: False 18:48:25 E 18:48:25 E 18:48:25 E (caused by) 18:48:25 E FunctionTestError: double backward is not implemented correctly 18:48:25 E 18:48:25 E (caused by) 18:48:25 E AssertionError: check_double_backward failed (eps=0.001 atol=0.001 rtol=0.001) ... 18:48:25 E gradients (numeric): -0.020542602051870057 18:48:25 E gradients (backward): -0.01935833355256278 18:48:25 E 18:48:25 E x: numeric gradient, y: backward gradient 18:48:25 E Not equal to tolerance rtol=0.001, atol=0.001 18:48:25 E 18:48:25 E (mismatch 100.0%) 18:48:25 E x: array(-0.020542602051870057) 18:48:25 E y: array(-0.01935833355256278) 18:48:25 E 18:48:25 E assert_allclose failed: 18:48:25 E shape: () () 18:48:25 E dtype: float64 float64 18:48:25 E i: (0,) 18:48:25 E x[i]: -0.020542602051870057 18:48:25 E y[i]: -0.01935833355256278 18:48:25 E relative error[i]: 0.061176159409160366 18:48:25 E absolute error[i]: 0.0011842684993072783 18:48:25 E relative tolerance * |y[i]|: 1.9358333552562778e-05 18:48:25 E absolute tolerance: 0.001 18:48:25 E total tolerance: 0.001019358333552563 18:48:25 E x: -0.020542602051870057 18:48:25 E y: -0.01935833355256278 ```
1.0
flaky test: `chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction::test_double_backward` - Possibly related to #7946 https://jenkins.preferred.jp/job/chainer/job/cupy_pr/347/TEST=chainer-py35,label=mn1-p100/console >`FAIL tests/chainer_tests/functions_tests/connection_tests/test_convolution_2d.py::TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always_param_25_{dilate=1, nobias=False, x_dtype=float16, contiguous=None, W_dtype=float16, groups=2, cover_all=False}::test_double_backward` ``` 18:48:25 E chainer.testing.function_link.FunctionTestError: Parameterized test failed. 18:48:25 E 18:48:25 E Base test method: TestConvolution2DFunction_use_chainerx_false__chainerx_device_None__use_cuda_false__cuda_device_None__use_cudnn_never__cudnn_deterministic_false__autotune_false__cudnn_fast_batch_normalization_false__use_ideep_always.test_double_backward 18:48:25 E Test parameters: 18:48:25 E dilate: 1 18:48:25 E nobias: False 18:48:25 E x_dtype: <class 'numpy.float16'> 18:48:25 E contiguous: None 18:48:25 E W_dtype: <class 'numpy.float16'> 18:48:25 E groups: 2 18:48:25 E cover_all: False 18:48:25 E 18:48:25 E 18:48:25 E (caused by) 18:48:25 E FunctionTestError: double backward is not implemented correctly 18:48:25 E 18:48:25 E (caused by) 18:48:25 E AssertionError: check_double_backward failed (eps=0.001 atol=0.001 rtol=0.001) ... 18:48:25 E gradients (numeric): -0.020542602051870057 18:48:25 E gradients (backward): -0.01935833355256278 18:48:25 E 18:48:25 E x: numeric gradient, y: backward gradient 18:48:25 E Not equal to tolerance rtol=0.001, atol=0.001 18:48:25 E 18:48:25 E (mismatch 100.0%) 18:48:25 E x: array(-0.020542602051870057) 18:48:25 E y: array(-0.01935833355256278) 18:48:25 E 18:48:25 E assert_allclose failed: 18:48:25 E shape: () () 18:48:25 E dtype: float64 float64 18:48:25 E i: (0,) 18:48:25 E x[i]: -0.020542602051870057 18:48:25 E y[i]: -0.01935833355256278 18:48:25 E relative error[i]: 0.061176159409160366 18:48:25 E absolute error[i]: 0.0011842684993072783 18:48:25 E relative tolerance * |y[i]|: 1.9358333552562778e-05 18:48:25 E absolute tolerance: 0.001 18:48:25 E total tolerance: 0.001019358333552563 18:48:25 E x: -0.020542602051870057 18:48:25 E y: -0.01935833355256278 ```
test
flaky test chainer tests functions tests connection tests test convolution py test double backward possibly related to fail tests chainer tests functions tests connection tests test convolution py use chainerx false chainerx device none use cuda false cuda device none use cudnn never cudnn deterministic false autotune false cudnn fast batch normalization false use ideep always param dilate nobias false x dtype contiguous none w dtype groups cover all false test double backward e chainer testing function link functiontesterror parameterized test failed e e base test method use chainerx false chainerx device none use cuda false cuda device none use cudnn never cudnn deterministic false autotune false cudnn fast batch normalization false use ideep always test double backward e test parameters e dilate e nobias false e x dtype e contiguous none e w dtype e groups e cover all false e e e caused by e functiontesterror double backward is not implemented correctly e e caused by e assertionerror check double backward failed eps atol rtol e gradients numeric e gradients backward e e x numeric gradient y backward gradient e not equal to tolerance rtol atol e e mismatch e x array e y array e e assert allclose failed e shape e dtype e i e x e y e relative error e absolute error e relative tolerance y e absolute tolerance e total tolerance e x e y
1
23,679
4,031,735,207
IssuesEvent
2016-05-18 18:11:03
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
circleci: failed tests: TestTxnPutOutOfOrder
Robot test-failure
The following test appears to have failed: [#18040](https://circleci.com/gh/cockroachdb/cockroach/18040): ``` I160518 17:35:26.740337 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740471 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740604 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740729 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740859 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> panic: test timed out after 1m10s goroutine 14692 [running]: panic(0x156f960, 0xc821c80050) /usr/local/go/src/runtime/panic.go:464 +0x3e6 testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:725 +0x14b created by time.goFunc /usr/local/go/src/time/sleep.go:129 +0x3a goroutine 1 [chan receive]: testing.RunTests(0x1d17878, 0x24a2540, 0x10f, 0x10f, 0xc820215b01) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc820045ef8, 0xc820265c98) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/storage_test.TestMain(0xc820045ef8) /go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de main.main() github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114 goroutine 17 [syscall, 1 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 20 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x277a0a0) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a goroutine 14674 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4a0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14635 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14640 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9f040) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14669 [select]: github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc820e94960) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm() /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d46800) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14638 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14637 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef80) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14678 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6a0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14675 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14690 [select]: github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc8215ff040) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14670 [select]: github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820eaede0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14665 [semacquire]: sync.runtime_Syncsemacquire(0xc821728f90) /usr/local/go/src/runtime/sema.go:241 +0x201 sync.(*Cond).Wait(0xc821728f80) /usr/local/go/src/sync/cond.go:63 +0x9b github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc821800310) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821800310) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7 github.com/cockroachdb/cockroach/storage_test.TestTxnPutOutOfOrder(0xc8219150e0) /go/src/github.com/cockroachdb/cockroach/storage/client_replica_test.go:322 +0xdaa testing.tRunner(0xc8219150e0, 0x24a3cb0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 14679 [select]: github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc821801260, 0xecece9eac, 0xaee4eb8, 0x27794c0, 0xc821728f40, 0xc821800310, 0xc821926300, 0xc820021e98) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc821926300, 0xc821224240) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78 github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821224690, 0xc821729e00) /go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14639 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efe0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14666 [chan receive]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82185dda0) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4 goroutine 14636 [runnable]: fmt.(*pp).argNumber(0xc821dd2ea0, 0x2, 0x1960640, 0xa, 0x8, 0x3, 0x1, 0x5, 0x0) /usr/local/go/src/fmt/print.go:1088 fmt.(*pp).doPrintf(0xc821dd2ea0, 0x1960640, 0xa, 0xc820914048, 0x3, 0x3) /usr/local/go/src/fmt/print.go:1144 +0xc87 fmt.Sprintf(0x1960640, 0xa, 0xc820914048, 0x3, 0x3, 0x0, 0x0) /usr/local/go/src/fmt/print.go:203 +0x6f github.com/cockroachdb/cockroach/roachpb.BatchRequest.String(0xc8, 0xdf388, 0x100000001, 0x1, 0x2, 0x0, 0x0, 0x0, 0xc8219d8ba0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/roachpb/batch.go:525 +0x64a github.com/cockroachdb/cockroach/roachpb.(*BatchRequest).String(0xc820c99340, 0x0, 0x0) <autogenerated>:414 +0xb3 -- github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef60) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14673 [select]: github.com/cockroachdb/cockroach/storage.(*Store).Start.func4() /go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc480) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14671 [select]: github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d47dc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 FAIL github.com/cockroachdb/cockroach/storage 70.049s === RUN TestBatchBasics I160518 17:34:06.434263 storage/engine/rocksdb.go:137 opening in memory rocksdb instance I160518 17:34:06.436714 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance --- PASS: TestBatchBasics (0.02s) === RUN TestBatchRepr I160518 17:34:06.441427 storage/engine/rocksdb.go:137 opening in memory rocksdb instance I160518 17:34:06.443054 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance --- PASS: TestBatchRepr (0.02s) === RUN TestBatchGet I160518 17:34:06.460742 storage/engine/rocksdb.go:137 opening in memory rocksdb instance ``` Please assign, take a look and update the issue accordingly.
1.0
circleci: failed tests: TestTxnPutOutOfOrder - The following test appears to have failed: [#18040](https://circleci.com/gh/cockroachdb/cockroach/18040): ``` I160518 17:35:26.740337 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740471 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740604 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740729 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> I160518 17:35:26.740859 kv/dist_sender.go:974 application error: range 0: replica <nil> not leader; leader is <nil> panic: test timed out after 1m10s goroutine 14692 [running]: panic(0x156f960, 0xc821c80050) /usr/local/go/src/runtime/panic.go:464 +0x3e6 testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:725 +0x14b created by time.goFunc /usr/local/go/src/time/sleep.go:129 +0x3a goroutine 1 [chan receive]: testing.RunTests(0x1d17878, 0x24a2540, 0x10f, 0x10f, 0xc820215b01) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc820045ef8, 0xc820265c98) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/storage_test.TestMain(0xc820045ef8) /go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de main.main() github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114 goroutine 17 [syscall, 1 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 20 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x277a0a0) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a goroutine 14674 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4a0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14635 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14640 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9f040) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14669 [select]: github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc820e94960) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm() /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d46800) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14638 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14637 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef80) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14678 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6a0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14675 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc4c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14690 [select]: github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc8215ff040) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14670 [select]: github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820eaede0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14665 [semacquire]: sync.runtime_Syncsemacquire(0xc821728f90) /usr/local/go/src/runtime/sema.go:241 +0x201 sync.(*Cond).Wait(0xc821728f80) /usr/local/go/src/sync/cond.go:63 +0x9b github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc821800310) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821800310) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7 github.com/cockroachdb/cockroach/storage_test.TestTxnPutOutOfOrder(0xc8219150e0) /go/src/github.com/cockroachdb/cockroach/storage/client_replica_test.go:322 +0xdaa testing.tRunner(0xc8219150e0, 0x24a3cb0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 14679 [select]: github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc821801260, 0xecece9eac, 0xaee4eb8, 0x27794c0, 0xc821728f40, 0xc821800310, 0xc821926300, 0xc820021e98) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc821926300, 0xc821224240) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78 github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821224690, 0xc821729e00) /go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc6c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14639 [select]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9efe0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14666 [chan receive]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82185dda0) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4 goroutine 14636 [runnable]: fmt.(*pp).argNumber(0xc821dd2ea0, 0x2, 0x1960640, 0xa, 0x8, 0x3, 0x1, 0x5, 0x0) /usr/local/go/src/fmt/print.go:1088 fmt.(*pp).doPrintf(0xc821dd2ea0, 0x1960640, 0xa, 0xc820914048, 0x3, 0x3) /usr/local/go/src/fmt/print.go:1144 +0xc87 fmt.Sprintf(0x1960640, 0xa, 0xc820914048, 0x3, 0x3, 0x0, 0x0) /usr/local/go/src/fmt/print.go:203 +0x6f github.com/cockroachdb/cockroach/roachpb.BatchRequest.String(0xc8, 0xdf388, 0x100000001, 0x1, 0x2, 0x0, 0x0, 0x0, 0xc8219d8ba0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/roachpb/batch.go:525 +0x64a github.com/cockroachdb/cockroach/roachpb.(*BatchRequest).String(0xc820c99340, 0x0, 0x0) <autogenerated>:414 +0xb3 -- github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820e9ef60) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14673 [select]: github.com/cockroachdb/cockroach/storage.(*Store).Start.func4() /go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820bbc480) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 14671 [select]: github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821800310, 0xc820d47dc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 FAIL github.com/cockroachdb/cockroach/storage 70.049s === RUN TestBatchBasics I160518 17:34:06.434263 storage/engine/rocksdb.go:137 opening in memory rocksdb instance I160518 17:34:06.436714 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance --- PASS: TestBatchBasics (0.02s) === RUN TestBatchRepr I160518 17:34:06.441427 storage/engine/rocksdb.go:137 opening in memory rocksdb instance I160518 17:34:06.443054 storage/engine/rocksdb.go:178 closing in-memory rocksdb instance --- PASS: TestBatchRepr (0.02s) === RUN TestBatchGet I160518 17:34:06.460742 storage/engine/rocksdb.go:137 opening in memory rocksdb instance ``` Please assign, take a look and update the issue accordingly.
test
circleci failed tests testtxnputoutoforder the following test appears to have failed kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is kv dist sender go application error range replica not leader leader is panic test timed out after goroutine panic usr local go src runtime panic go testing startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach storage test testmain go src github com cockroachdb cockroach storage main test go main main github com cockroachdb cockroach storage test testmain go goroutine runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach util log loggingt flushdaemon go src github com cockroachdb cockroach util log clog go created by github com cockroachdb cockroach util log init go src github com cockroachdb cockroach util log clog go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach kv txncoordsender startstats go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach kv txncoordsender github com cockroachdb cockroach kv startstats fm go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage idallocator start go src github com cockroachdb cockroach storage id alloc go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage storepool start go src github com cockroachdb cockroach storage store pool go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine sync runtime syncsemacquire usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach util stop stopper quiesce go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach util stop stopper stop go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach storage test testtxnputoutoforder go src github com cockroachdb cockroach storage client replica test go testing trunner usr local go src testing testing go created by testing runtests usr local go src testing testing go goroutine github com cockroachdb cockroach storage replicascanner waitandprocess go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage storerangeset visit go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go created by github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go goroutine fmt pp argnumber usr local go src fmt print go fmt pp doprintf usr local go src fmt print go fmt sprintf usr local go src fmt print go github com cockroachdb cockroach roachpb batchrequest string go src github com cockroachdb cockroach roachpb batch go github com cockroachdb cockroach roachpb batchrequest string github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store start go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store processraft go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go fail github com cockroachdb cockroach storage run testbatchbasics storage engine rocksdb go opening in memory rocksdb instance storage engine rocksdb go closing in memory rocksdb instance pass testbatchbasics run testbatchrepr storage engine rocksdb go opening in memory rocksdb instance storage engine rocksdb go closing in memory rocksdb instance pass testbatchrepr run testbatchget storage engine rocksdb go opening in memory rocksdb instance please assign take a look and update the issue accordingly
1
209,017
15,961,986,764
IssuesEvent
2021-04-16 00:12:37
Carlosma7/TFM-GrandQuiz
https://api.github.com/repos/Carlosma7/TFM-GrandQuiz
closed
Definición tests clase Partida
test
Se deben definir los tests asociados a la clase Partida definida en #46 .
1.0
Definición tests clase Partida - Se deben definir los tests asociados a la clase Partida definida en #46 .
test
definición tests clase partida se deben definir los tests asociados a la clase partida definida en
1
286,420
8,788,060,385
IssuesEvent
2018-12-20 20:48:58
RPGHacker/asar
https://api.github.com/repos/RPGHacker/asar
closed
Segfault with defines and quotes and stuff
bug priority: medium
``` !w = "lol""lol" print "!w" ``` Segfaults. The 2nd line gets parsed to `print "lol"lol"` after define evaluation, which causes qsplit to throw up.
1.0
Segfault with defines and quotes and stuff - ``` !w = "lol""lol" print "!w" ``` Segfaults. The 2nd line gets parsed to `print "lol"lol"` after define evaluation, which causes qsplit to throw up.
non_test
segfault with defines and quotes and stuff w lol lol print w segfaults the line gets parsed to print lol lol after define evaluation which causes qsplit to throw up
0
136,566
18,741,071,817
IssuesEvent
2021-11-04 13:40:26
ioana-github-enterprise/testremediate
https://api.github.com/repos/ioana-github-enterprise/testremediate
opened
CVE-2021-32769 (High) detected in micronaut-inject-2.3.1.jar, micronaut-core-2.3.1.jar
security vulnerability
## CVE-2021-32769 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>micronaut-inject-2.3.1.jar</b>, <b>micronaut-core-2.3.1.jar</b></p></summary> <p> <details><summary><b>micronaut-inject-2.3.1.jar</b></p></summary> <p>Natively Cloud Native</p> <p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p> <p>Path to vulnerable library: /micronaut-inject-2.3.1.jar</p> <p> Dependency Hierarchy: - :x: **micronaut-inject-2.3.1.jar** (Vulnerable Library) </details> <details><summary><b>micronaut-core-2.3.1.jar</b></p></summary> <p>Natively Cloud Native</p> <p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p> <p>Path to vulnerable library: /micronaut-core-2.3.1.jar</p> <p> Dependency Hierarchy: - :x: **micronaut-core-2.3.1.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/testremediate/commit/c50f50f9e3a35342bd1f045ceb06f33a49e525fb">c50f50f9e3a35342bd1f045ceb06f33a49e525fb</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using "/../../" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot. <p>Publish Date: 2021-07-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769>CVE-2021-32769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769</a></p> <p>Release Date: 2021-07-16</p> <p>Fix Resolution: io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-inject","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-inject:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"},{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-core","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-core:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32769","vulnerabilityDetails":"Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using \"/../../\" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-32769 (High) detected in micronaut-inject-2.3.1.jar, micronaut-core-2.3.1.jar - ## CVE-2021-32769 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>micronaut-inject-2.3.1.jar</b>, <b>micronaut-core-2.3.1.jar</b></p></summary> <p> <details><summary><b>micronaut-inject-2.3.1.jar</b></p></summary> <p>Natively Cloud Native</p> <p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p> <p>Path to vulnerable library: /micronaut-inject-2.3.1.jar</p> <p> Dependency Hierarchy: - :x: **micronaut-inject-2.3.1.jar** (Vulnerable Library) </details> <details><summary><b>micronaut-core-2.3.1.jar</b></p></summary> <p>Natively Cloud Native</p> <p>Library home page: <a href="http://micronaut.io">http://micronaut.io</a></p> <p>Path to vulnerable library: /micronaut-core-2.3.1.jar</p> <p> Dependency Hierarchy: - :x: **micronaut-core-2.3.1.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/testremediate/commit/c50f50f9e3a35342bd1f045ceb06f33a49e525fb">c50f50f9e3a35342bd1f045ceb06f33a49e525fb</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using "/../../" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot. <p>Publish Date: 2021-07-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769>CVE-2021-32769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32769</a></p> <p>Release Date: 2021-07-16</p> <p>Fix Resolution: io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-inject","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-inject:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"},{"packageType":"Java","groupId":"io.micronaut","packageName":"micronaut-core","packageVersion":"2.3.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"io.micronaut:micronaut-core:2.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.micronaut:micronaut-core:2.5.9, io.micronaut:micronaut-inject:2.5.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-32769","vulnerabilityDetails":"Micronaut is a JVM-based, full stack Java framework designed for building JVM applications. A path traversal vulnerability exists in versions prior to 2.5.9. With a basic configuration, it is possible to access any file from a filesystem, using \"/../../\" in the URL. This occurs because Micronaut does not restrict file access to configured paths. The vulnerability is patched in version 2.5.9. As a workaround, do not use `**` in mapping, use only `*`, which exposes only flat structure of a directory not allowing traversal. If using Linux, another workaround is to run micronaut in chroot.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32769","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in micronaut inject jar micronaut core jar cve high severity vulnerability vulnerable libraries micronaut inject jar micronaut core jar micronaut inject jar natively cloud native library home page a href path to vulnerable library micronaut inject jar dependency hierarchy x micronaut inject jar vulnerable library micronaut core jar natively cloud native library home page a href path to vulnerable library micronaut core jar dependency hierarchy x micronaut core jar vulnerable library found in head commit a href found in base branch main vulnerability details micronaut is a jvm based full stack java framework designed for building jvm applications a path traversal vulnerability exists in versions prior to with a basic configuration it is possible to access any file from a filesystem using in the url this occurs because micronaut does not restrict file access to configured paths the vulnerability is patched in version as a workaround do not use in mapping use only which exposes only flat structure of a directory not allowing traversal if using linux another workaround is to run micronaut in chroot publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io micronaut micronaut core io micronaut micronaut inject isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree io micronaut micronaut inject isminimumfixversionavailable true minimumfixversion io micronaut micronaut core io micronaut micronaut inject packagetype java groupid io micronaut packagename micronaut core packageversion packagefilepaths istransitivedependency false dependencytree io micronaut micronaut core isminimumfixversionavailable true minimumfixversion io micronaut micronaut core io micronaut micronaut inject basebranches vulnerabilityidentifier cve vulnerabilitydetails micronaut is a jvm based full stack java framework designed for building jvm applications a path traversal vulnerability exists in versions prior to with a basic configuration it is possible to access any file from a filesystem using in the url this occurs because micronaut does not restrict file access to configured paths the vulnerability is patched in version as a workaround do not use in mapping use only which exposes only flat structure of a directory not allowing traversal if using linux another workaround is to run micronaut in chroot vulnerabilityurl
0
85,653
3,697,239,299
IssuesEvent
2016-02-27 14:53:44
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Creating HPA using docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml fails validation
area/autoscaling kind/bug priority/P1 team/control-plane
I ran: ``` $ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml ``` which gave: ``` error validating "docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml": error validating data: found invalid field namespace for v1beta1.SubresourceReference; if you choose to ignore these errors, turn validation off with --validate=false ``` Adding validate=false ran fine. ``` $ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml --validate=false ``` I was running HEAD code locally. cc @piosz @kubernetes/autoscaling
1.0
Creating HPA using docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml fails validation - I ran: ``` $ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml ``` which gave: ``` error validating "docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml": error validating data: found invalid field namespace for v1beta1.SubresourceReference; if you choose to ignore these errors, turn validation off with --validate=false ``` Adding validate=false ran fine. ``` $ ./cluster/kubectl.sh create -f docs/user-guide/horizontal-pod-autoscaling/hpa-php-apache.yaml --validate=false ``` I was running HEAD code locally. cc @piosz @kubernetes/autoscaling
non_test
creating hpa using docs user guide horizontal pod autoscaling hpa php apache yaml fails validation i ran cluster kubectl sh create f docs user guide horizontal pod autoscaling hpa php apache yaml which gave error validating docs user guide horizontal pod autoscaling hpa php apache yaml error validating data found invalid field namespace for subresourcereference if you choose to ignore these errors turn validation off with validate false adding validate false ran fine cluster kubectl sh create f docs user guide horizontal pod autoscaling hpa php apache yaml validate false i was running head code locally cc piosz kubernetes autoscaling
0
56,812
6,529,126,406
IssuesEvent
2017-08-30 10:16:01
edenlabllc/ehealth.api
https://api.github.com/repos/edenlabllc/ehealth.api
closed
Prevent employee update via employee request for DISMISSED employee
kind/bug priority/high status/test
Prevent employee update via employee request for DISMISSED employee Return 409 with message "DISMISSED employee can't be updated" **Request:** ``` curl -X POST \ http://dev.ehealth.world/api/employee_requests \ -H 'authorization: Bearer bVp1b1M4TGEzNWcvMWFMa1RmRW5iUT09' \ -H 'cache-control: no-cache' \ -H 'content-type: application/json' \ -H 'postman-token: d391baef-b66c-0fb3-c561-baec99a19f03' \ -d '{ "employee_request": { "legal_entity_id": "f41ebb33-e9d0-4ef4-8e92-ee719d16554e", "employee_id":"33afca77-9e1e-4c79-8a5a-fc8164b4da85", "position": "P1", "start_date": "2017-03-02", "status": "NEW", "employee_type": "DOCTOR", "party": { "first_name": "Петро", "last_name": "Іванов", "second_name": "Миколайович", "birth_date": "1991-08-25", "gender": "MALE", "tax_id": "3067305998", "email": "sp.virny+51@gmail.com", "documents": [ { "type": "PASSPORT", "number": "120518" } ], "phones": [ { "type": "MOBILE", "number": "+380503410870" } ] }, "doctor": { "educations": [ { "country": "UA", "city": "Київ", "institution_name": "Академія Богомольця", "issued_date": "2017-08-05", "diploma_number": "DD123543", "degree": "MASTER", "speciality": "Педіатр" } ], "qualifications": [ { "type": "CLINICAL_RESIDENCY", "institution_name": "Академія Богомольця", "speciality": "Педіатр", "issued_date": "2017-08-05", "certificate_number": "2017-08-05" } ], "specialities": [ { "speciality": "PEDIATRICIAN", "speciality_officio": true, "level": "FIRST", "qualification_type": "AWARDING", "attestation_name": "Академія Богомольця", "attestation_date": "2017-08-05", "valid_to_date": "2017-08-05", "certificate_number": "AB/21331" } ], "science_degree": { "country": "UA", "city": "Київ", "degree": "PhD", "institution_name": "Академія Богомольця", "diploma_number": "DD123543", "speciality": "THERAPIST", "issued_date": "2017-08-05" } } } }' ``` **Actual result:** 200 Success **Expected result:** 409 with message "DISMISSED employee can't be updated"
1.0
Prevent employee update via employee request for DISMISSED employee - Prevent employee update via employee request for DISMISSED employee Return 409 with message "DISMISSED employee can't be updated" **Request:** ``` curl -X POST \ http://dev.ehealth.world/api/employee_requests \ -H 'authorization: Bearer bVp1b1M4TGEzNWcvMWFMa1RmRW5iUT09' \ -H 'cache-control: no-cache' \ -H 'content-type: application/json' \ -H 'postman-token: d391baef-b66c-0fb3-c561-baec99a19f03' \ -d '{ "employee_request": { "legal_entity_id": "f41ebb33-e9d0-4ef4-8e92-ee719d16554e", "employee_id":"33afca77-9e1e-4c79-8a5a-fc8164b4da85", "position": "P1", "start_date": "2017-03-02", "status": "NEW", "employee_type": "DOCTOR", "party": { "first_name": "Петро", "last_name": "Іванов", "second_name": "Миколайович", "birth_date": "1991-08-25", "gender": "MALE", "tax_id": "3067305998", "email": "sp.virny+51@gmail.com", "documents": [ { "type": "PASSPORT", "number": "120518" } ], "phones": [ { "type": "MOBILE", "number": "+380503410870" } ] }, "doctor": { "educations": [ { "country": "UA", "city": "Київ", "institution_name": "Академія Богомольця", "issued_date": "2017-08-05", "diploma_number": "DD123543", "degree": "MASTER", "speciality": "Педіатр" } ], "qualifications": [ { "type": "CLINICAL_RESIDENCY", "institution_name": "Академія Богомольця", "speciality": "Педіатр", "issued_date": "2017-08-05", "certificate_number": "2017-08-05" } ], "specialities": [ { "speciality": "PEDIATRICIAN", "speciality_officio": true, "level": "FIRST", "qualification_type": "AWARDING", "attestation_name": "Академія Богомольця", "attestation_date": "2017-08-05", "valid_to_date": "2017-08-05", "certificate_number": "AB/21331" } ], "science_degree": { "country": "UA", "city": "Київ", "degree": "PhD", "institution_name": "Академія Богомольця", "diploma_number": "DD123543", "speciality": "THERAPIST", "issued_date": "2017-08-05" } } } }' ``` **Actual result:** 200 Success **Expected result:** 409 with message "DISMISSED employee can't be updated"
test
prevent employee update via employee request for dismissed employee prevent employee update via employee request for dismissed employee return with message dismissed employee can t be updated request curl x post h authorization bearer h cache control no cache h content type application json h postman token d employee request legal entity id employee id position start date status new employee type doctor party first name петро last name іванов second name миколайович birth date gender male tax id email sp virny gmail com documents type passport number phones type mobile number doctor educations country ua city київ institution name академія богомольця issued date diploma number degree master speciality педіатр qualifications type clinical residency institution name академія богомольця speciality педіатр issued date certificate number specialities speciality pediatrician speciality officio true level first qualification type awarding attestation name академія богомольця attestation date valid to date certificate number ab science degree country ua city київ degree phd institution name академія богомольця diploma number speciality therapist issued date actual result success expected result with message dismissed employee can t be updated
1
591,359
17,837,887,088
IssuesEvent
2021-09-03 05:41:51
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
minikube start fails with podman or kvm2 drivers on s390x
kind/feature priority/backlog
minikube claims to run on s390x, but I have not had success yet. I am on Fedora 33 here, so perhaps another disto would would better, but I've seen Ubuntu fail similarly using the kvm2 driver. <!--- Please include the "minikube start" command you used in your reproduction steps ---> **Steps to reproduce the issue:** 1. minikube config set driver podman && minikube delete 2. minikube start *or* 1. minikube config set driver kvm2 && minikube delete 2. minikube start **Full output of `minikube logs` command (for `podman`)** <details> ==> Last Start <== Log file created at: 2021/06/15 14:42:08 Running on machine: minikube1 Binary: Built with gc go1.16.4 for linux/s390x Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg I0615 14:42:08.728959 30552 out.go:291] Setting OutFile to fd 1 ... I0615 14:42:08.729059 30552 out.go:343] isatty.IsTerminal(1) = true I0615 14:42:08.729067 30552 out.go:304] Setting ErrFile to fd 2... I0615 14:42:08.729071 30552 out.go:343] isatty.IsTerminal(2) = true I0615 14:42:08.729176 30552 root.go:316] Updating PATH: /home/fedora/.minikube/bin I0615 14:42:08.729372 30552 out.go:298] Setting JSON to false I0615 14:42:08.730027 30552 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3644,"bootTime":1623764485,"procs":351,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"} I0615 14:42:08.730074 30552 start.go:121] virtualization: kvm host I0615 14:42:08.732896 30552 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x) I0615 14:42:08.733072 30552 notify.go:169] Checking for updates... I0615 14:42:08.733074 30552 driver.go:335] Setting default libvirt URI to qemu:///system I0615 14:42:08.893533 30552 podman.go:121] podman version: 3.1.2 I0615 14:42:08.894674 30552 out.go:170] ✨ Using the podman driver based on user configuration I0615 14:42:08.894726 30552 start.go:279] selected driver: podman I0615 14:42:08.894729 30552 start.go:752] validating driver "podman" against <nil> I0615 14:42:08.894736 30552 start.go:763] status for podman: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc:} I0615 14:42:08.894791 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:09.043747 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1246912512 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 43.41s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:09.043798 30552 start_flags.go:259] no existing cluster config was found, will generate one from the flags I0615 14:42:09.043940 30552 start_flags.go:311] Using suggested 2200MB memory alloc based on sys=4004MB, container=4004MB I0615 14:42:09.044002 30552 start_flags.go:638] Wait components to verify : map[apiserver:true system_pods:true] I0615 14:42:09.044013 30552 cni.go:93] Creating CNI manager for "" I0615 14:42:09.044019 30552 cni.go:167] CNI unnecessary in this configuration, recommending no CNI I0615 14:42:09.044022 30552 start_flags.go:273] config: {Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} I0615 14:42:09.045663 30552 out.go:170] 👍 Starting control plane node minikube in cluster minikube I0615 14:42:09.045679 30552 cache.go:115] Beginning downloading kic base image for podman with docker I0615 14:42:09.046691 30552 out.go:170] 🚜 Pulling base image ... I0615 14:42:09.046719 30552 preload.go:110] Checking if preload exists for k8s version v1.20.7 and runtime docker I0615 14:42:09.046784 30552 cache.go:134] Downloading gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache I0615 14:42:09.046910 30552 image.go:58] Checking for gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 in local cache directory I0615 14:42:09.046927 30552 image.go:118] Writing gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache W0615 14:42:09.106869 30552 preload.go:140] https://storage.googleapis.com/minikube-preloaded-volume-tarballs/preloaded-images-k8s-v11-v1.20.7-docker-overlay2-s390x.tar.lz4 status code: 404 I0615 14:42:09.107061 30552 profile.go:148] Saving config to /home/fedora/.minikube/profiles/minikube/config.json ... I0615 14:42:09.107073 30552 lock.go:36] WriteFile acquiring /home/fedora/.minikube/profiles/minikube/config.json: {Name:mkb8d756e7e807b96e4ac95488558156b33e3a18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>} I0615 14:42:09.107211 30552 cache.go:108] acquiring lock: {Name:mk848f9056a0b7e8deba61c4d9ce1ec9447fb24f Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107243 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 exists I0615 14:42:09.107250 30552 cache.go:97] cache image "docker.io/kubernetesui/metrics-scraper:v1.0.4" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4" took 40.935µs I0615 14:42:09.107256 30552 cache.go:81] save to tar file docker.io/kubernetesui/metrics-scraper:v1.0.4 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 succeeded I0615 14:42:09.107265 30552 cache.go:108] acquiring lock: {Name:mke06d93f1117b77439f99f030a503df9447235e Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107293 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 exists I0615 14:42:09.107299 30552 cache.go:97] cache image "k8s.gcr.io/kube-apiserver:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7" took 35.361µs I0615 14:42:09.107304 30552 cache.go:81] save to tar file k8s.gcr.io/kube-apiserver:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 succeeded I0615 14:42:09.107312 30552 cache.go:108] acquiring lock: {Name:mk2f2a385b1cfc81a9ca9ce72bba6acd289ea36a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107339 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 exists I0615 14:42:09.107344 30552 cache.go:97] cache image "k8s.gcr.io/kube-controller-manager:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7" took 33.518µs I0615 14:42:09.107349 30552 cache.go:81] save to tar file k8s.gcr.io/kube-controller-manager:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 succeeded I0615 14:42:09.107357 30552 cache.go:108] acquiring lock: {Name:mkbf90ca757cecf35c48d78ea90ab30ec4c88b47 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107383 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 exists I0615 14:42:09.107388 30552 cache.go:97] cache image "k8s.gcr.io/kube-scheduler:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7" took 32.516µs I0615 14:42:09.107393 30552 cache.go:81] save to tar file k8s.gcr.io/kube-scheduler:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 succeeded I0615 14:42:09.107400 30552 cache.go:108] acquiring lock: {Name:mk42495b17f7ffd0cee7bc305c566b5bc1271eab Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107426 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 exists I0615 14:42:09.107432 30552 cache.go:97] cache image "k8s.gcr.io/kube-proxy:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7" took 32.595µs I0615 14:42:09.107437 30552 cache.go:81] save to tar file k8s.gcr.io/kube-proxy:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 succeeded I0615 14:42:09.107445 30552 cache.go:108] acquiring lock: {Name:mk15a5e7d0f8daa4a4ee292d980fdb92a25be656 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107470 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 exists I0615 14:42:09.107476 30552 cache.go:97] cache image "k8s.gcr.io/pause:3.2" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2" took 32.821µs I0615 14:42:09.107480 30552 cache.go:81] save to tar file k8s.gcr.io/pause:3.2 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 succeeded I0615 14:42:09.107488 30552 cache.go:108] acquiring lock: {Name:mke705c7f34a678272a870eb3ee742e676627306 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107513 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 exists I0615 14:42:09.107519 30552 cache.go:97] cache image "k8s.gcr.io/etcd:3.4.13-0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0" took 32.046µs I0615 14:42:09.107523 30552 cache.go:81] save to tar file k8s.gcr.io/etcd:3.4.13-0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 succeeded I0615 14:42:09.107531 30552 cache.go:108] acquiring lock: {Name:mk58b9291a6bc97c1f1359b79ec84521b6f848fc Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107557 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 exists I0615 14:42:09.107563 30552 cache.go:97] cache image "k8s.gcr.io/coredns:1.7.0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0" took 33.11µs I0615 14:42:09.107567 30552 cache.go:81] save to tar file k8s.gcr.io/coredns:1.7.0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 succeeded I0615 14:42:09.107575 30552 cache.go:108] acquiring lock: {Name:mk92ab5604f143a7cec05887d019b5678bdb0226 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107602 30552 cache.go:116] /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 exists I0615 14:42:09.107607 30552 cache.go:97] cache image "gcr.io/k8s-minikube/storage-provisioner:v5" -> "/home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5" took 33.748µs I0615 14:42:09.107613 30552 cache.go:81] save to tar file gcr.io/k8s-minikube/storage-provisioner:v5 -> /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 succeeded I0615 14:42:09.107620 30552 cache.go:108] acquiring lock: {Name:mk9e35de248c46aa08df049e2c423012df463dc5 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107645 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 exists I0615 14:42:09.107651 30552 cache.go:97] cache image "docker.io/kubernetesui/dashboard:v2.1.0" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0" took 31.152µs I0615 14:42:09.107655 30552 cache.go:81] save to tar file docker.io/kubernetesui/dashboard:v2.1.0 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 succeeded I0615 14:42:09.107658 30552 cache.go:88] Successfully saved all images to host disk. E0615 14:42:09.283440 30552 cache.go:197] Error downloading kic artifacts: not yet implemented, see issue #8426 I0615 14:42:09.283450 30552 cache.go:202] Successfully downloaded all kic artifacts I0615 14:42:09.283462 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.283495 30552 start.go:317] acquired machines lock for "minikube" in 27.196µs I0615 14:42:09.283504 30552 start.go:89] Provisioning new machine with config: &{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} &{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true} I0615 14:42:09.283536 30552 start.go:126] createHost starting for "" (driver="podman") I0615 14:42:09.285059 30552 out.go:197] 🔥 Creating podman container (CPUs=2, Memory=2200MB) ... I0615 14:42:09.285218 30552 start.go:160] libmachine.API.Create for "minikube" (driver="podman") I0615 14:42:09.285230 30552 client.go:168] LocalClient.Create starting I0615 14:42:09.285267 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/ca.pem I0615 14:42:09.285286 30552 main.go:128] libmachine: Decoding PEM data... I0615 14:42:09.285297 30552 main.go:128] libmachine: Parsing certificate... I0615 14:42:09.285368 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/cert.pem I0615 14:42:09.285384 30552 main.go:128] libmachine: Decoding PEM data... I0615 14:42:09.285393 30552 main.go:128] libmachine: Parsing certificate... I0615 14:42:09.285627 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube --format "{{range .plugins}}{{if eq .type "bridge"}}{{(index (index .ipam.ranges 0) 0).subnet}},{{(index (index .ipam.ranges 0) 0).gateway}}{{end}}{{end}}" I0615 14:42:09.393427 30552 network_create.go:67] Found existing network {name:minikube subnet:0xc000b75920 gateway:[0 0 0 0 0 0 0 0 0 0 255 255 192 168 49 1] mtu:0} I0615 14:42:09.393438 30552 kic.go:106] calculated static IP "192.168.49.2" for the "minikube" container I0615 14:42:09.393480 30552 cli_runner.go:115] Run: sudo -n podman ps -a --format {{.Names}} I0615 14:42:09.523500 30552 cli_runner.go:115] Run: sudo -n podman volume create minikube --label name.minikube.sigs.k8s.io=minikube --label created_by.minikube.sigs.k8s.io=true I0615 14:42:09.653396 30552 oci.go:102] Successfully created a podman volume minikube I0615 14:42:09.653434 30552 cli_runner.go:115] Run: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib W0615 14:42:10.133503 30552 cli_runner.go:162] sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib returned with exit code 125 I0615 14:42:10.133523 30552 client.go:171] LocalClient.Create took 848.289535ms I0615 14:42:12.134090 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'" I0615 14:42:12.134129 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:12.243575 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:12.353526 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:12.353575 30552 retry.go:31] will retry after 276.165072ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:12.630208 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:12.733566 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:12.843522 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:12.843573 30552 retry.go:31] will retry after 540.190908ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:13.384633 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:13.513535 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:13.623653 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:13.623710 30552 retry.go:31] will retry after 655.06503ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:14.279839 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:14.423423 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:14.543373 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 W0615 14:42:14.543426 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:14.543432 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:14.543436 30552 start.go:129] duration metric: createHost completed in 5.259896732s I0615 14:42:14.543440 30552 start.go:80] releasing machines lock for "minikube", held for 5.259942055s W0615 14:42:14.543449 30552 start.go:518] error starting host: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125 stdout: stderr: Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23... no image found in manifest list for architecture s390x, variant "", OS linux Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux I0615 14:42:14.543773 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}} W0615 14:42:14.683483 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125 I0615 14:42:14.683504 30552 delete.go:82] Unable to get host status for minikube, assuming it has already been deleted: state: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:14.683568 30552 out.go:235] 🤦 StartHost failed, but will try again: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125 stdout: stderr: Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23... no image found in manifest list for architecture s390x, variant "", OS linux Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux I0615 14:42:14.683719 30552 start.go:533] Will try again in 5 seconds ... I0615 14:42:19.684326 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:19.684380 30552 start.go:317] acquired machines lock for "minikube" in 42.164µs I0615 14:42:19.684389 30552 start.go:93] Skipping create...Using existing machine configuration I0615 14:42:19.684393 30552 fix.go:55] fixHost starting: I0615 14:42:19.684563 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}} W0615 14:42:19.813413 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125 I0615 14:42:19.813429 30552 fix.go:108] recreateIfNeeded on minikube: state= err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:19.813439 30552 fix.go:113] machineExists: true. err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:19.813444 30552 fix.go:134] unexpected machine state, will restart: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:19.815104 30552 out.go:170] 🔄 Restarting existing podman container for "minikube" ... I0615 14:42:19.815145 30552 cli_runner.go:115] Run: sudo -n podman start --cgroup-manager cgroupfs minikube W0615 14:42:19.953379 30552 cli_runner.go:162] sudo -n podman start --cgroup-manager cgroupfs minikube returned with exit code 125 I0615 14:42:19.953414 30552 cli_runner.go:115] Run: sudo -n podman inspect minikube I0615 14:42:20.073749 30552 errors.go:84] Postmortem inspect ("sudo -n podman inspect minikube"): -- stdout -- [ { "Name": "minikube", "Driver": "local", "Mountpoint": "/var/lib/containers/storage/volumes/minikube/_data", "CreatedAt": "2021-06-15T14:42:09.614053773Z", "Labels": { "created_by.minikube.sigs.k8s.io": "true", "name.minikube.sigs.k8s.io": "minikube" }, "Scope": "local", "Options": {} } ] -- /stdout -- I0615 14:42:20.073838 30552 cli_runner.go:115] Run: sudo -n podman logs --timestamps minikube W0615 14:42:20.203437 30552 cli_runner.go:162] sudo -n podman logs --timestamps minikube returned with exit code 125 W0615 14:42:20.203447 30552 errors.go:89] Failed to get postmortem logs. sudo -n podman logs --timestamps minikube :sudo -n podman logs --timestamps minikube: exit status 125 stdout: stderr: Error: no container with name or ID "minikube" found: no such container I0615 14:42:20.203478 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:20.353543 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.353562 30552 errors.go:106] postmortem podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.353590 30552 network_create.go:255] running [podman network inspect minikube] to gather additional debugging logs... I0615 14:42:20.353609 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube I0615 14:42:20.483538 30552 network_create.go:260] output of [sudo -n podman network inspect minikube]: -- stdout -- [ { "cniVersion": "0.4.0", "name": "minikube", "plugins": [ { "bridge": "cni-podman1", "hairpinMode": true, "ipMasq": true, "ipam": { "ranges": [ [ { "gateway": "192.168.49.1", "subnet": "192.168.49.0/24" } ] ], "routes": [ { "dst": "0.0.0.0/0" } ], "type": "host-local" }, "isGateway": true, "type": "bridge" }, { "capabilities": { "portMappings": true }, "type": "portmap" }, { "backend": "", "type": "firewall" }, { "type": "tuning" }, { "capabilities": { "aliases": true }, "domainName": "dns.podman", "type": "dnsname" } ] } ] -- /stdout -- I0615 14:42:20.483583 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:20.643548 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243791360 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 55.01s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.643777 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube W0615 14:42:20.793518 30552 cli_runner.go:162] sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube returned with exit code 125 I0615 14:42:20.793559 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'" I0615 14:42:20.793592 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:20.943495 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:21.083399 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:21.083457 30552 retry.go:31] will retry after 234.428547ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:21.318877 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:21.443589 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:21.573394 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:21.573444 30552 retry.go:31] will retry after 346.739061ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:21.920931 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:22.053534 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:22.203360 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:22.203409 30552 retry.go:31] will retry after 667.587979ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:22.871144 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:22.993505 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:23.123357 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 W0615 14:42:23.123417 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:23.123423 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:23.123428 30552 fix.go:57] fixHost completed within 3.439034957s I0615 14:42:23.123432 30552 start.go:80] releasing machines lock for "minikube", held for 3.43904856s W0615 14:42:23.123519 30552 out.go:235] 😿 Failed to start podman container. Running "minikube delete" may fix it: podman inspect ip minikube: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:23.125747 30552 out.go:170] W0615 14:42:23.125806 30552 out.go:235] ❌ Exiting due to GUEST_PROVISION: Failed to start host: podman inspect ip minikube: sudo -n podman container inspect -f minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:23.125913 30552 out.go:235] W0615 14:42:23.126777 30552 out.go:235] ╭────────────────────────────────────────────────────────────────────╮ W0615 14:42:23.126788 30552 out.go:235] │ │ W0615 14:42:23.126807 30552 out.go:235] │ 😿 If the above advice does not help, please let us know: │ W0615 14:42:23.126827 30552 out.go:235] │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ W0615 14:42:23.126842 30552 out.go:235] │ │ W0615 14:42:23.126856 30552 out.go:235] │ Please attach the following file to the GitHub issue: │ W0615 14:42:23.126867 30552 out.go:235] │ - /home/fedora/.minikube/logs/lastStart.txt │ W0615 14:42:23.126877 30552 out.go:235] │ │ W0615 14:42:23.126888 30552 out.go:235] ╰────────────────────────────────────────────────────────────────────╯ W0615 14:42:23.126900 30552 out.go:235] ❌ Exiting due to GUEST_STATUS: state: unknown state "minikube": sudo -n podman container inspect minikube --format=: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube ╭───────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 If the above advice does not help, please let us know: │ │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ │ │ │ Please attach the following file to the GitHub issue: │ │ - /tmp/minikube_logs_f30b94c7b8be27a1785d74f9772c624a74c09c39_0.log │ │ │ ╰───────────────────────────────────────────────────────────────────────────╯ </details> **Full output of `minikube logs` command (for `kvm2`)** <details> ==> Audit <== |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| | Command | Args | Profile | User | Version | Start Time | End Time | |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:05 UTC | Tue, 15 Jun 2021 13:57:05 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:13 UTC | Tue, 15 Jun 2021 13:57:13 UTC | | start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:38 UTC | Tue, 15 Jun 2021 14:29:38 UTC | | start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:49 UTC | Tue, 15 Jun 2021 14:29:49 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:35 UTC | Tue, 15 Jun 2021 14:37:35 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:40 UTC | Tue, 15 Jun 2021 14:37:40 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:18 UTC | Tue, 15 Jun 2021 14:40:18 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC | | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:06 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:07 UTC | | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC | |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| ==> Last Start <== Log file created at: 2021/06/15 14:45:09 Running on machine: minikube1 Binary: Built with gc go1.16.4 for linux/s390x Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg I0615 14:45:09.494355 35758 out.go:291] Setting OutFile to fd 1 ... I0615 14:45:09.494462 35758 out.go:343] isatty.IsTerminal(1) = true I0615 14:45:09.494465 35758 out.go:304] Setting ErrFile to fd 2... I0615 14:45:09.494468 35758 out.go:343] isatty.IsTerminal(2) = true I0615 14:45:09.494545 35758 root.go:316] Updating PATH: /home/fedora/.minikube/bin I0615 14:45:09.494725 35758 out.go:298] Setting JSON to false I0615 14:45:09.495146 35758 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3824,"bootTime":1623764485,"procs":92,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"} I0615 14:45:09.495190 35758 start.go:121] virtualization: kvm host I0615 14:45:09.496967 35758 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x) I0615 14:45:09.497089 35758 notify.go:169] Checking for updates... I0615 14:45:09.497486 35758 driver.go:335] Setting default libvirt URI to qemu:///system I0615 14:45:09.498712 35758 out.go:170] ✨ Using the kvm2 driver based on user configuration I0615 14:45:09.498722 35758 start.go:279] selected driver: kvm2 I0615 14:45:09.498725 35758 start.go:752] validating driver "kvm2" against <nil> I0615 14:45:09.498731 35758 start.go:763] status for kvm2: {Installed:true Healthy:false Running:true NeedsImprovement:false Error:libvirt group membership check failed: error getting current user's GIDs: user: GroupIds requires cgo Reason:PR_KVM_USER_PERMISSION Fix:Check that libvirtd is properly installed and that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) Doc:https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/} I0615 14:45:09.499768 35758 out.go:170] W0615 14:45:09.499816 35758 out.go:235] 🚫 Exiting due to PR_KVM_USER_PERMISSION: libvirt group membership check failed: error getting current user's GIDs: user: GroupIds requires cgo W0615 14:45:09.499939 35758 out.go:235] 💡 Suggestion: Ensure that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) W0615 14:45:09.499974 35758 out.go:235] 📘 Documentation: https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/ W0615 14:45:09.499985 35758 out.go:235] 🍿 Related issues: W0615 14:45:09.500018 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/5617 W0615 14:45:09.500049 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/10070 🤷 Profile "minikube" not found. Run "minikube profile list" to view all profiles. 👉 To start a cluster, run: "minikube start" </details>
1.0
minikube start fails with podman or kvm2 drivers on s390x - minikube claims to run on s390x, but I have not had success yet. I am on Fedora 33 here, so perhaps another disto would would better, but I've seen Ubuntu fail similarly using the kvm2 driver. <!--- Please include the "minikube start" command you used in your reproduction steps ---> **Steps to reproduce the issue:** 1. minikube config set driver podman && minikube delete 2. minikube start *or* 1. minikube config set driver kvm2 && minikube delete 2. minikube start **Full output of `minikube logs` command (for `podman`)** <details> ==> Last Start <== Log file created at: 2021/06/15 14:42:08 Running on machine: minikube1 Binary: Built with gc go1.16.4 for linux/s390x Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg I0615 14:42:08.728959 30552 out.go:291] Setting OutFile to fd 1 ... I0615 14:42:08.729059 30552 out.go:343] isatty.IsTerminal(1) = true I0615 14:42:08.729067 30552 out.go:304] Setting ErrFile to fd 2... I0615 14:42:08.729071 30552 out.go:343] isatty.IsTerminal(2) = true I0615 14:42:08.729176 30552 root.go:316] Updating PATH: /home/fedora/.minikube/bin I0615 14:42:08.729372 30552 out.go:298] Setting JSON to false I0615 14:42:08.730027 30552 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3644,"bootTime":1623764485,"procs":351,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"} I0615 14:42:08.730074 30552 start.go:121] virtualization: kvm host I0615 14:42:08.732896 30552 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x) I0615 14:42:08.733072 30552 notify.go:169] Checking for updates... I0615 14:42:08.733074 30552 driver.go:335] Setting default libvirt URI to qemu:///system I0615 14:42:08.893533 30552 podman.go:121] podman version: 3.1.2 I0615 14:42:08.894674 30552 out.go:170] ✨ Using the podman driver based on user configuration I0615 14:42:08.894726 30552 start.go:279] selected driver: podman I0615 14:42:08.894729 30552 start.go:752] validating driver "podman" against <nil> I0615 14:42:08.894736 30552 start.go:763] status for podman: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc:} I0615 14:42:08.894791 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:09.043747 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1246912512 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 43.41s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:09.043798 30552 start_flags.go:259] no existing cluster config was found, will generate one from the flags I0615 14:42:09.043940 30552 start_flags.go:311] Using suggested 2200MB memory alloc based on sys=4004MB, container=4004MB I0615 14:42:09.044002 30552 start_flags.go:638] Wait components to verify : map[apiserver:true system_pods:true] I0615 14:42:09.044013 30552 cni.go:93] Creating CNI manager for "" I0615 14:42:09.044019 30552 cni.go:167] CNI unnecessary in this configuration, recommending no CNI I0615 14:42:09.044022 30552 start_flags.go:273] config: {Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} I0615 14:42:09.045663 30552 out.go:170] 👍 Starting control plane node minikube in cluster minikube I0615 14:42:09.045679 30552 cache.go:115] Beginning downloading kic base image for podman with docker I0615 14:42:09.046691 30552 out.go:170] 🚜 Pulling base image ... I0615 14:42:09.046719 30552 preload.go:110] Checking if preload exists for k8s version v1.20.7 and runtime docker I0615 14:42:09.046784 30552 cache.go:134] Downloading gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache I0615 14:42:09.046910 30552 image.go:58] Checking for gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 in local cache directory I0615 14:42:09.046927 30552 image.go:118] Writing gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 to local cache W0615 14:42:09.106869 30552 preload.go:140] https://storage.googleapis.com/minikube-preloaded-volume-tarballs/preloaded-images-k8s-v11-v1.20.7-docker-overlay2-s390x.tar.lz4 status code: 404 I0615 14:42:09.107061 30552 profile.go:148] Saving config to /home/fedora/.minikube/profiles/minikube/config.json ... I0615 14:42:09.107073 30552 lock.go:36] WriteFile acquiring /home/fedora/.minikube/profiles/minikube/config.json: {Name:mkb8d756e7e807b96e4ac95488558156b33e3a18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>} I0615 14:42:09.107211 30552 cache.go:108] acquiring lock: {Name:mk848f9056a0b7e8deba61c4d9ce1ec9447fb24f Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107243 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 exists I0615 14:42:09.107250 30552 cache.go:97] cache image "docker.io/kubernetesui/metrics-scraper:v1.0.4" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4" took 40.935µs I0615 14:42:09.107256 30552 cache.go:81] save to tar file docker.io/kubernetesui/metrics-scraper:v1.0.4 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/metrics-scraper_v1.0.4 succeeded I0615 14:42:09.107265 30552 cache.go:108] acquiring lock: {Name:mke06d93f1117b77439f99f030a503df9447235e Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107293 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 exists I0615 14:42:09.107299 30552 cache.go:97] cache image "k8s.gcr.io/kube-apiserver:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7" took 35.361µs I0615 14:42:09.107304 30552 cache.go:81] save to tar file k8s.gcr.io/kube-apiserver:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-apiserver_v1.20.7 succeeded I0615 14:42:09.107312 30552 cache.go:108] acquiring lock: {Name:mk2f2a385b1cfc81a9ca9ce72bba6acd289ea36a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107339 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 exists I0615 14:42:09.107344 30552 cache.go:97] cache image "k8s.gcr.io/kube-controller-manager:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7" took 33.518µs I0615 14:42:09.107349 30552 cache.go:81] save to tar file k8s.gcr.io/kube-controller-manager:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-controller-manager_v1.20.7 succeeded I0615 14:42:09.107357 30552 cache.go:108] acquiring lock: {Name:mkbf90ca757cecf35c48d78ea90ab30ec4c88b47 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107383 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 exists I0615 14:42:09.107388 30552 cache.go:97] cache image "k8s.gcr.io/kube-scheduler:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7" took 32.516µs I0615 14:42:09.107393 30552 cache.go:81] save to tar file k8s.gcr.io/kube-scheduler:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-scheduler_v1.20.7 succeeded I0615 14:42:09.107400 30552 cache.go:108] acquiring lock: {Name:mk42495b17f7ffd0cee7bc305c566b5bc1271eab Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107426 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 exists I0615 14:42:09.107432 30552 cache.go:97] cache image "k8s.gcr.io/kube-proxy:v1.20.7" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7" took 32.595µs I0615 14:42:09.107437 30552 cache.go:81] save to tar file k8s.gcr.io/kube-proxy:v1.20.7 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.20.7 succeeded I0615 14:42:09.107445 30552 cache.go:108] acquiring lock: {Name:mk15a5e7d0f8daa4a4ee292d980fdb92a25be656 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107470 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 exists I0615 14:42:09.107476 30552 cache.go:97] cache image "k8s.gcr.io/pause:3.2" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2" took 32.821µs I0615 14:42:09.107480 30552 cache.go:81] save to tar file k8s.gcr.io/pause:3.2 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/pause_3.2 succeeded I0615 14:42:09.107488 30552 cache.go:108] acquiring lock: {Name:mke705c7f34a678272a870eb3ee742e676627306 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107513 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 exists I0615 14:42:09.107519 30552 cache.go:97] cache image "k8s.gcr.io/etcd:3.4.13-0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0" took 32.046µs I0615 14:42:09.107523 30552 cache.go:81] save to tar file k8s.gcr.io/etcd:3.4.13-0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/etcd_3.4.13-0 succeeded I0615 14:42:09.107531 30552 cache.go:108] acquiring lock: {Name:mk58b9291a6bc97c1f1359b79ec84521b6f848fc Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107557 30552 cache.go:116] /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 exists I0615 14:42:09.107563 30552 cache.go:97] cache image "k8s.gcr.io/coredns:1.7.0" -> "/home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0" took 33.11µs I0615 14:42:09.107567 30552 cache.go:81] save to tar file k8s.gcr.io/coredns:1.7.0 -> /home/fedora/.minikube/cache/images/k8s.gcr.io/coredns_1.7.0 succeeded I0615 14:42:09.107575 30552 cache.go:108] acquiring lock: {Name:mk92ab5604f143a7cec05887d019b5678bdb0226 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107602 30552 cache.go:116] /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 exists I0615 14:42:09.107607 30552 cache.go:97] cache image "gcr.io/k8s-minikube/storage-provisioner:v5" -> "/home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5" took 33.748µs I0615 14:42:09.107613 30552 cache.go:81] save to tar file gcr.io/k8s-minikube/storage-provisioner:v5 -> /home/fedora/.minikube/cache/images/gcr.io/k8s-minikube/storage-provisioner_v5 succeeded I0615 14:42:09.107620 30552 cache.go:108] acquiring lock: {Name:mk9e35de248c46aa08df049e2c423012df463dc5 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.107645 30552 cache.go:116] /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 exists I0615 14:42:09.107651 30552 cache.go:97] cache image "docker.io/kubernetesui/dashboard:v2.1.0" -> "/home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0" took 31.152µs I0615 14:42:09.107655 30552 cache.go:81] save to tar file docker.io/kubernetesui/dashboard:v2.1.0 -> /home/fedora/.minikube/cache/images/docker.io/kubernetesui/dashboard_v2.1.0 succeeded I0615 14:42:09.107658 30552 cache.go:88] Successfully saved all images to host disk. E0615 14:42:09.283440 30552 cache.go:197] Error downloading kic artifacts: not yet implemented, see issue #8426 I0615 14:42:09.283450 30552 cache.go:202] Successfully downloaded all kic artifacts I0615 14:42:09.283462 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:09.283495 30552 start.go:317] acquired machines lock for "minikube" in 27.196µs I0615 14:42:09.283504 30552 start.go:89] Provisioning new machine with config: &{Name:minikube KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.23@sha256:baf6d94b2050bcbecd98994e265cf965a4f4768978620ccf5227a6dcb75ade45 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:podman HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.20.7 ClusterName:minikube Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin: FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: MultiNodeRequested:false} &{Name: IP: Port:8443 KubernetesVersion:v1.20.7 ControlPlane:true Worker:true} I0615 14:42:09.283536 30552 start.go:126] createHost starting for "" (driver="podman") I0615 14:42:09.285059 30552 out.go:197] 🔥 Creating podman container (CPUs=2, Memory=2200MB) ... I0615 14:42:09.285218 30552 start.go:160] libmachine.API.Create for "minikube" (driver="podman") I0615 14:42:09.285230 30552 client.go:168] LocalClient.Create starting I0615 14:42:09.285267 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/ca.pem I0615 14:42:09.285286 30552 main.go:128] libmachine: Decoding PEM data... I0615 14:42:09.285297 30552 main.go:128] libmachine: Parsing certificate... I0615 14:42:09.285368 30552 main.go:128] libmachine: Reading certificate data from /home/fedora/.minikube/certs/cert.pem I0615 14:42:09.285384 30552 main.go:128] libmachine: Decoding PEM data... I0615 14:42:09.285393 30552 main.go:128] libmachine: Parsing certificate... I0615 14:42:09.285627 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube --format "{{range .plugins}}{{if eq .type "bridge"}}{{(index (index .ipam.ranges 0) 0).subnet}},{{(index (index .ipam.ranges 0) 0).gateway}}{{end}}{{end}}" I0615 14:42:09.393427 30552 network_create.go:67] Found existing network {name:minikube subnet:0xc000b75920 gateway:[0 0 0 0 0 0 0 0 0 0 255 255 192 168 49 1] mtu:0} I0615 14:42:09.393438 30552 kic.go:106] calculated static IP "192.168.49.2" for the "minikube" container I0615 14:42:09.393480 30552 cli_runner.go:115] Run: sudo -n podman ps -a --format {{.Names}} I0615 14:42:09.523500 30552 cli_runner.go:115] Run: sudo -n podman volume create minikube --label name.minikube.sigs.k8s.io=minikube --label created_by.minikube.sigs.k8s.io=true I0615 14:42:09.653396 30552 oci.go:102] Successfully created a podman volume minikube I0615 14:42:09.653434 30552 cli_runner.go:115] Run: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib W0615 14:42:10.133503 30552 cli_runner.go:162] sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib returned with exit code 125 I0615 14:42:10.133523 30552 client.go:171] LocalClient.Create took 848.289535ms I0615 14:42:12.134090 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'" I0615 14:42:12.134129 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:12.243575 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:12.353526 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:12.353575 30552 retry.go:31] will retry after 276.165072ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:12.630208 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:12.733566 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:12.843522 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:12.843573 30552 retry.go:31] will retry after 540.190908ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:13.384633 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:13.513535 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:13.623653 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:13.623710 30552 retry.go:31] will retry after 655.06503ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:14.279839 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:14.423423 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:14.543373 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 W0615 14:42:14.543426 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:14.543432 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:14.543436 30552 start.go:129] duration metric: createHost completed in 5.259896732s I0615 14:42:14.543440 30552 start.go:80] releasing machines lock for "minikube", held for 5.259942055s W0615 14:42:14.543449 30552 start.go:518] error starting host: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125 stdout: stderr: Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23... no image found in manifest list for architecture s390x, variant "", OS linux Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux I0615 14:42:14.543773 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}} W0615 14:42:14.683483 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125 I0615 14:42:14.683504 30552 delete.go:82] Unable to get host status for minikube, assuming it has already been deleted: state: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:14.683568 30552 out.go:235] 🤦 StartHost failed, but will try again: creating host: create: creating: setting up container node: preparing volume for minikube container: sudo -n podman run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.23 -d /var/lib: exit status 125 stdout: stderr: Trying to pull gcr.io/k8s-minikube/kicbase:v0.0.23... no image found in manifest list for architecture s390x, variant "", OS linux Error: Error choosing an image from manifest list docker://gcr.io/k8s-minikube/kicbase:v0.0.23: no image found in manifest list for architecture s390x, variant "", OS linux I0615 14:42:14.683719 30552 start.go:533] Will try again in 5 seconds ... I0615 14:42:19.684326 30552 start.go:313] acquiring machines lock for minikube: {Name:mkfbd64e670de175ef3ec6dd8be25ea1851f8d07 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>} I0615 14:42:19.684380 30552 start.go:317] acquired machines lock for "minikube" in 42.164µs I0615 14:42:19.684389 30552 start.go:93] Skipping create...Using existing machine configuration I0615 14:42:19.684393 30552 fix.go:55] fixHost starting: I0615 14:42:19.684563 30552 cli_runner.go:115] Run: sudo -n podman container inspect minikube --format={{.State.Status}} W0615 14:42:19.813413 30552 cli_runner.go:162] sudo -n podman container inspect minikube --format={{.State.Status}} returned with exit code 125 I0615 14:42:19.813429 30552 fix.go:108] recreateIfNeeded on minikube: state= err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:19.813439 30552 fix.go:113] machineExists: true. err=unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:19.813444 30552 fix.go:134] unexpected machine state, will restart: unknown state "minikube": sudo -n podman container inspect minikube --format={{.State.Status}}: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:19.815104 30552 out.go:170] 🔄 Restarting existing podman container for "minikube" ... I0615 14:42:19.815145 30552 cli_runner.go:115] Run: sudo -n podman start --cgroup-manager cgroupfs minikube W0615 14:42:19.953379 30552 cli_runner.go:162] sudo -n podman start --cgroup-manager cgroupfs minikube returned with exit code 125 I0615 14:42:19.953414 30552 cli_runner.go:115] Run: sudo -n podman inspect minikube I0615 14:42:20.073749 30552 errors.go:84] Postmortem inspect ("sudo -n podman inspect minikube"): -- stdout -- [ { "Name": "minikube", "Driver": "local", "Mountpoint": "/var/lib/containers/storage/volumes/minikube/_data", "CreatedAt": "2021-06-15T14:42:09.614053773Z", "Labels": { "created_by.minikube.sigs.k8s.io": "true", "name.minikube.sigs.k8s.io": "minikube" }, "Scope": "local", "Options": {} } ] -- /stdout -- I0615 14:42:20.073838 30552 cli_runner.go:115] Run: sudo -n podman logs --timestamps minikube W0615 14:42:20.203437 30552 cli_runner.go:162] sudo -n podman logs --timestamps minikube returned with exit code 125 W0615 14:42:20.203447 30552 errors.go:89] Failed to get postmortem logs. sudo -n podman logs --timestamps minikube :sudo -n podman logs --timestamps minikube: exit status 125 stdout: stderr: Error: no container with name or ID "minikube" found: no such container I0615 14:42:20.203478 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:20.353543 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.353562 30552 errors.go:106] postmortem podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243226112 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 54.72s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.353590 30552 network_create.go:255] running [podman network inspect minikube] to gather additional debugging logs... I0615 14:42:20.353609 30552 cli_runner.go:115] Run: sudo -n podman network inspect minikube I0615 14:42:20.483538 30552 network_create.go:260] output of [sudo -n podman network inspect minikube]: -- stdout -- [ { "cniVersion": "0.4.0", "name": "minikube", "plugins": [ { "bridge": "cni-podman1", "hairpinMode": true, "ipMasq": true, "ipam": { "ranges": [ [ { "gateway": "192.168.49.1", "subnet": "192.168.49.0/24" } ] ], "routes": [ { "dst": "0.0.0.0/0" } ], "type": "host-local" }, "isGateway": true, "type": "bridge" }, { "capabilities": { "portMappings": true }, "type": "portmap" }, { "backend": "", "type": "firewall" }, { "type": "tuning" }, { "capabilities": { "aliases": true }, "domainName": "dns.podman", "type": "dnsname" } ] } ] -- /stdout -- I0615 14:42:20.483583 30552 cli_runner.go:115] Run: sudo -n podman system info --format json I0615 14:42:20.643548 30552 info.go:281] podman info: {Host:{BuildahVersion:1.20.1 CgroupVersion:v2 Conmon:{Package:conmon-2.0.27-2.fc33.s390x Path:/usr/bin/conmon Version:conmon version 2.0.27, commit: } Distribution:{Distribution:fedora Version:33} MemFree:1243791360 MemTotal:4198801408 OCIRuntime:{Name:crun Package:crun-0.19.1-3.fc33.s390x Path:/usr/bin/crun Version:crun version 0.19.1 commit: 1535fedf0b83fb898d449f9680000f729ba719f5 spec: 1.0.0 +SYSTEMD +SELINUX +APPARMOR +CAP +SECCOMP +EBPF +CRIU +YAJL} SwapFree:0 SwapTotal:0 Arch:s390x Cpus:2 Eventlogger:journald Hostname:minikube1.zdalisv.dfw.ibm.com Kernel:5.12.10-200.fc33.s390x Os:linux Rootless:false Uptime:1h 0m 55.01s (Approximately 0.04 days)} Registries:{Search:[registry.fedoraproject.org registry.access.redhat.com docker.io]} Store:{ConfigFile:/etc/containers/storage.conf ContainerStore:{Number:0} GraphDriverName:overlay GraphOptions:{} GraphRoot:/var/lib/containers/storage GraphStatus:{BackingFilesystem:extfs NativeOverlayDiff:true SupportsDType:true UsingMetacopy:false} ImageStore:{Number:0} RunRoot:/run/containers/storage VolumePath:/var/lib/containers/storage/volumes}} I0615 14:42:20.643777 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube W0615 14:42:20.793518 30552 cli_runner.go:162] sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube returned with exit code 125 I0615 14:42:20.793559 30552 ssh_runner.go:149] Run: sh -c "df -h /var | awk 'NR==2{print $5}'" I0615 14:42:20.793592 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:20.943495 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:21.083399 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:21.083457 30552 retry.go:31] will retry after 234.428547ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:21.318877 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:21.443589 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:21.573394 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:21.573444 30552 retry.go:31] will retry after 346.739061ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:21.920931 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:22.053534 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:22.203360 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 I0615 14:42:22.203409 30552 retry.go:31] will retry after 667.587979ms: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:22.871144 30552 cli_runner.go:115] Run: sudo -n podman version --format {{.Version}} I0615 14:42:22.993505 30552 cli_runner.go:115] Run: sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube W0615 14:42:23.123357 30552 cli_runner.go:162] sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube returned with exit code 125 W0615 14:42:23.123417 30552 start.go:257] error running df -h /var: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:23.123423 30552 start.go:239] error getting percentage of /var that is free: NewSession: new client: new client: Error creating new ssh host from driver: Error getting ssh port for driver: get ssh host-port: get port 22 for "minikube": sudo -n podman container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:23.123428 30552 fix.go:57] fixHost completed within 3.439034957s I0615 14:42:23.123432 30552 start.go:80] releasing machines lock for "minikube", held for 3.43904856s W0615 14:42:23.123519 30552 out.go:235] 😿 Failed to start podman container. Running "minikube delete" may fix it: podman inspect ip minikube: sudo -n podman container inspect -f {{.NetworkSettings.IPAddress}} minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube I0615 14:42:23.125747 30552 out.go:170] W0615 14:42:23.125806 30552 out.go:235] ❌ Exiting due to GUEST_PROVISION: Failed to start host: podman inspect ip minikube: sudo -n podman container inspect -f minikube: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube W0615 14:42:23.125913 30552 out.go:235] W0615 14:42:23.126777 30552 out.go:235] ╭────────────────────────────────────────────────────────────────────╮ W0615 14:42:23.126788 30552 out.go:235] │ │ W0615 14:42:23.126807 30552 out.go:235] │ 😿 If the above advice does not help, please let us know: │ W0615 14:42:23.126827 30552 out.go:235] │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ W0615 14:42:23.126842 30552 out.go:235] │ │ W0615 14:42:23.126856 30552 out.go:235] │ Please attach the following file to the GitHub issue: │ W0615 14:42:23.126867 30552 out.go:235] │ - /home/fedora/.minikube/logs/lastStart.txt │ W0615 14:42:23.126877 30552 out.go:235] │ │ W0615 14:42:23.126888 30552 out.go:235] ╰────────────────────────────────────────────────────────────────────╯ W0615 14:42:23.126900 30552 out.go:235] ❌ Exiting due to GUEST_STATUS: state: unknown state "minikube": sudo -n podman container inspect minikube --format=: exit status 125 stdout: stderr: Error: error inspecting object: no such container minikube ╭───────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 If the above advice does not help, please let us know: │ │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ │ │ │ Please attach the following file to the GitHub issue: │ │ - /tmp/minikube_logs_f30b94c7b8be27a1785d74f9772c624a74c09c39_0.log │ │ │ ╰───────────────────────────────────────────────────────────────────────────╯ </details> **Full output of `minikube logs` command (for `kvm2`)** <details> ==> Audit <== |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| | Command | Args | Profile | User | Version | Start Time | End Time | |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:05 UTC | Tue, 15 Jun 2021 13:57:05 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 13:57:13 UTC | Tue, 15 Jun 2021 13:57:13 UTC | | start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:38 UTC | Tue, 15 Jun 2021 14:29:38 UTC | | start | --help | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:29:49 UTC | Tue, 15 Jun 2021 14:29:49 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:35 UTC | Tue, 15 Jun 2021 14:37:35 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:37:40 UTC | Tue, 15 Jun 2021 14:37:40 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:18 UTC | Tue, 15 Jun 2021 14:40:18 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:22 UTC | Tue, 15 Jun 2021 14:40:22 UTC | | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:40:56 UTC | Tue, 15 Jun 2021 14:40:56 UTC | | config | set driver podman | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:06 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:42:06 UTC | Tue, 15 Jun 2021 14:42:07 UTC | | config | set driver kvm2 | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC | | delete | | minikube | fedora | v1.21.0 | Tue, 15 Jun 2021 14:45:05 UTC | Tue, 15 Jun 2021 14:45:05 UTC | |---------|-------------------|----------|--------|---------|-------------------------------|-------------------------------| ==> Last Start <== Log file created at: 2021/06/15 14:45:09 Running on machine: minikube1 Binary: Built with gc go1.16.4 for linux/s390x Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg I0615 14:45:09.494355 35758 out.go:291] Setting OutFile to fd 1 ... I0615 14:45:09.494462 35758 out.go:343] isatty.IsTerminal(1) = true I0615 14:45:09.494465 35758 out.go:304] Setting ErrFile to fd 2... I0615 14:45:09.494468 35758 out.go:343] isatty.IsTerminal(2) = true I0615 14:45:09.494545 35758 root.go:316] Updating PATH: /home/fedora/.minikube/bin I0615 14:45:09.494725 35758 out.go:298] Setting JSON to false I0615 14:45:09.495146 35758 start.go:111] hostinfo: {"hostname":"minikube1.zdalisv.dfw.ibm.com","uptime":3824,"bootTime":1623764485,"procs":92,"os":"linux","platform":"fedora","platformFamily":"fedora","platformVersion":"33","kernelVersion":"5.12.10-200.fc33.s390x","kernelArch":"s390x","virtualizationSystem":"kvm","virtualizationRole":"host","hostId":"457c81ed-72d0-436d-a743-6a916ac685bb"} I0615 14:45:09.495190 35758 start.go:121] virtualization: kvm host I0615 14:45:09.496967 35758 out.go:170] 😄 minikube v1.21.0 on Fedora 33 (s390x) I0615 14:45:09.497089 35758 notify.go:169] Checking for updates... I0615 14:45:09.497486 35758 driver.go:335] Setting default libvirt URI to qemu:///system I0615 14:45:09.498712 35758 out.go:170] ✨ Using the kvm2 driver based on user configuration I0615 14:45:09.498722 35758 start.go:279] selected driver: kvm2 I0615 14:45:09.498725 35758 start.go:752] validating driver "kvm2" against <nil> I0615 14:45:09.498731 35758 start.go:763] status for kvm2: {Installed:true Healthy:false Running:true NeedsImprovement:false Error:libvirt group membership check failed: error getting current user's GIDs: user: GroupIds requires cgo Reason:PR_KVM_USER_PERMISSION Fix:Check that libvirtd is properly installed and that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) Doc:https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/} I0615 14:45:09.499768 35758 out.go:170] W0615 14:45:09.499816 35758 out.go:235] 🚫 Exiting due to PR_KVM_USER_PERMISSION: libvirt group membership check failed: error getting current user's GIDs: user: GroupIds requires cgo W0615 14:45:09.499939 35758 out.go:235] 💡 Suggestion: Ensure that you are a member of the appropriate libvirt group (remember to relogin for group changes to take effect!) W0615 14:45:09.499974 35758 out.go:235] 📘 Documentation: https://minikube.sigs.k8s.io/docs/reference/drivers/kvm2/ W0615 14:45:09.499985 35758 out.go:235] 🍿 Related issues: W0615 14:45:09.500018 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/5617 W0615 14:45:09.500049 35758 out.go:235] ▪ https://github.com/kubernetes/minikube/issues/10070 🤷 Profile "minikube" not found. Run "minikube profile list" to view all profiles. 👉 To start a cluster, run: "minikube start" </details>
non_test
minikube start fails with podman or drivers on minikube claims to run on but i have not had success yet i am on fedora here so perhaps another disto would would better but i ve seen ubuntu fail similarly using the driver steps to reproduce the issue minikube config set driver podman minikube delete minikube start or minikube config set driver minikube delete minikube start full output of minikube logs command for podman last start log file created at running on machine binary built with gc for linux log line format mmdd hh mm ss uuuuuu threadid file line msg out go setting outfile to fd out go isatty isterminal true out go setting errfile to fd out go isatty isterminal true root go updating path home fedora minikube bin out go setting json to false start go hostinfo hostname zdalisv dfw ibm com uptime boottime procs os linux platform fedora platformfamily fedora platformversion kernelversion kernelarch virtualizationsystem kvm virtualizationrole host hostid start go virtualization kvm host out go 😄 minikube on fedora notify go checking for updates driver go setting default libvirt uri to qemu system podman go podman version out go ✨ using the podman driver based on user configuration start go selected driver podman start go validating driver podman against start go status for podman installed true healthy true running false needsimprovement false error reason fix doc cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes start flags go no existing cluster config was found will generate one from the flags start flags go using suggested memory alloc based on sys container start flags go wait components to verify map cni go creating cni manager for cni go cni unnecessary in this configuration recommending no cni start flags go config name minikube keepcontext false embedcerts false minikubeiso kicbaseimage gcr io minikube kicbase memory cpus disksize vmdriver driver podman hyperkitvpnkitsock hyperkitvsockports dockerenv containervolumemounts insecureregistry registrymirror hostonlycidr hypervvirtualswitch hypervuseexternalswitch false hypervexternaladapter kvmnetwork default kvmqemuuri qemu system kvmgpu false kvmhidden false kvmnumacount dockeropt disabledrivermounts false nfsshare nfssharesroot nfsshares uuid novtxcheck false dnsproxy false hostdnsresolver true hostonlynictype virtio natnictype virtio sshipaddress sshuser root sshkey sshport kubernetesconfig kubernetesversion clustername minikube namespace default apiservername minikubeca apiservernames apiserverips dnsdomain cluster local containerruntime docker crisocket networkplugin featuregates servicecidr imagerepository loadbalancerstartip loadbalancerendip customingresscert extraoptions shouldloadcachedimages true enabledefaultcni false cni nodeip nodeport nodename nodes addons map customaddonimages map customaddonregistries map verifycomponents map starthosttimeout scheduledstop exposedports listenaddress network multinoderequested false out go 👍 starting control plane node minikube in cluster minikube cache go beginning downloading kic base image for podman with docker out go 🚜 pulling base image preload go checking if preload exists for version and runtime docker cache go downloading gcr io minikube kicbase to local cache image go checking for gcr io minikube kicbase in local cache directory image go writing gcr io minikube kicbase to local cache preload go status code profile go saving config to home fedora minikube profiles minikube config json lock go writefile acquiring home fedora minikube profiles minikube config json name clock delay timeout cancel cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images docker io kubernetesui metrics scraper exists cache go cache image docker io kubernetesui metrics scraper home fedora minikube cache images docker io kubernetesui metrics scraper took cache go save to tar file docker io kubernetesui metrics scraper home fedora minikube cache images docker io kubernetesui metrics scraper succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube apiserver exists cache go cache image gcr io kube apiserver home fedora minikube cache images gcr io kube apiserver took cache go save to tar file gcr io kube apiserver home fedora minikube cache images gcr io kube apiserver succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube controller manager exists cache go cache image gcr io kube controller manager home fedora minikube cache images gcr io kube controller manager took cache go save to tar file gcr io kube controller manager home fedora minikube cache images gcr io kube controller manager succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube scheduler exists cache go cache image gcr io kube scheduler home fedora minikube cache images gcr io kube scheduler took cache go save to tar file gcr io kube scheduler home fedora minikube cache images gcr io kube scheduler succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io kube proxy exists cache go cache image gcr io kube proxy home fedora minikube cache images gcr io kube proxy took cache go save to tar file gcr io kube proxy home fedora minikube cache images gcr io kube proxy succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io pause exists cache go cache image gcr io pause home fedora minikube cache images gcr io pause took cache go save to tar file gcr io pause home fedora minikube cache images gcr io pause succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io etcd exists cache go cache image gcr io etcd home fedora minikube cache images gcr io etcd took cache go save to tar file gcr io etcd home fedora minikube cache images gcr io etcd succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io coredns exists cache go cache image gcr io coredns home fedora minikube cache images gcr io coredns took cache go save to tar file gcr io coredns home fedora minikube cache images gcr io coredns succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images gcr io minikube storage provisioner exists cache go cache image gcr io minikube storage provisioner home fedora minikube cache images gcr io minikube storage provisioner took cache go save to tar file gcr io minikube storage provisioner home fedora minikube cache images gcr io minikube storage provisioner succeeded cache go acquiring lock name clock delay timeout cancel cache go home fedora minikube cache images docker io kubernetesui dashboard exists cache go cache image docker io kubernetesui dashboard home fedora minikube cache images docker io kubernetesui dashboard took cache go save to tar file docker io kubernetesui dashboard home fedora minikube cache images docker io kubernetesui dashboard succeeded cache go successfully saved all images to host disk cache go error downloading kic artifacts not yet implemented see issue cache go successfully downloaded all kic artifacts start go acquiring machines lock for minikube name clock delay timeout cancel start go acquired machines lock for minikube in start go provisioning new machine with config name minikube keepcontext false embedcerts false minikubeiso kicbaseimage gcr io minikube kicbase memory cpus disksize vmdriver driver podman hyperkitvpnkitsock hyperkitvsockports dockerenv containervolumemounts insecureregistry registrymirror hostonlycidr hypervvirtualswitch hypervuseexternalswitch false hypervexternaladapter kvmnetwork default kvmqemuuri qemu system kvmgpu false kvmhidden false kvmnumacount dockeropt disabledrivermounts false nfsshare nfssharesroot nfsshares uuid novtxcheck false dnsproxy false hostdnsresolver true hostonlynictype virtio natnictype virtio sshipaddress sshuser root sshkey sshport kubernetesconfig kubernetesversion clustername minikube namespace default apiservername minikubeca apiservernames apiserverips dnsdomain cluster local containerruntime docker crisocket networkplugin featuregates servicecidr imagerepository loadbalancerstartip loadbalancerendip customingresscert extraoptions shouldloadcachedimages true enabledefaultcni false cni nodeip nodeport nodename nodes addons map customaddonimages map customaddonregistries map verifycomponents map starthosttimeout scheduledstop exposedports listenaddress network multinoderequested false name ip port kubernetesversion controlplane true worker true start go createhost starting for driver podman out go 🔥 creating podman container cpus memory start go libmachine api create for minikube driver podman client go localclient create starting main go libmachine reading certificate data from home fedora minikube certs ca pem main go libmachine decoding pem data main go libmachine parsing certificate main go libmachine reading certificate data from home fedora minikube certs cert pem main go libmachine decoding pem data main go libmachine parsing certificate cli runner go run sudo n podman network inspect minikube format range plugins if eq type bridge index index ipam ranges subnet index index ipam ranges gateway end end network create go found existing network name minikube subnet gateway mtu kic go calculated static ip for the minikube container cli runner go run sudo n podman ps a format names cli runner go run sudo n podman volume create minikube label name minikube sigs io minikube label created by minikube sigs io true oci go successfully created a podman volume minikube cli runner go run sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib cli runner go sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib returned with exit code client go localclient create took ssh runner go run sh c df h var awk nr print cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code start go error running df h var newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go error getting percentage of var that is free newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go duration metric createhost completed in start go releasing machines lock for minikube held for start go error starting host creating host create creating setting up container node preparing volume for minikube container sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib exit status stdout stderr trying to pull gcr io minikube kicbase no image found in manifest list for architecture variant os linux error error choosing an image from manifest list docker gcr io minikube kicbase no image found in manifest list for architecture variant os linux cli runner go run sudo n podman container inspect minikube format state status cli runner go sudo n podman container inspect minikube format state status returned with exit code delete go unable to get host status for minikube assuming it has already been deleted state unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube out go 🤦 starthost failed but will try again creating host create creating setting up container node preparing volume for minikube container sudo n podman run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib exit status stdout stderr trying to pull gcr io minikube kicbase no image found in manifest list for architecture variant os linux error error choosing an image from manifest list docker gcr io minikube kicbase no image found in manifest list for architecture variant os linux start go will try again in seconds start go acquiring machines lock for minikube name clock delay timeout cancel start go acquired machines lock for minikube in start go skipping create using existing machine configuration fix go fixhost starting cli runner go run sudo n podman container inspect minikube format state status cli runner go sudo n podman container inspect minikube format state status returned with exit code fix go recreateifneeded on minikube state err unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube fix go machineexists true err unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube fix go unexpected machine state will restart unknown state minikube sudo n podman container inspect minikube format state status exit status stdout stderr error error inspecting object no such container minikube out go 🔄 restarting existing podman container for minikube cli runner go run sudo n podman start cgroup manager cgroupfs minikube cli runner go sudo n podman start cgroup manager cgroupfs minikube returned with exit code cli runner go run sudo n podman inspect minikube errors go postmortem inspect sudo n podman inspect minikube stdout name minikube driver local mountpoint var lib containers storage volumes minikube data createdat labels created by minikube sigs io true name minikube sigs io minikube scope local options stdout cli runner go run sudo n podman logs timestamps minikube cli runner go sudo n podman logs timestamps minikube returned with exit code errors go failed to get postmortem logs sudo n podman logs timestamps minikube sudo n podman logs timestamps minikube exit status stdout stderr error no container with name or id minikube found no such container cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes errors go postmortem podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes network create go running to gather additional debugging logs cli runner go run sudo n podman network inspect minikube network create go output of stdout cniversion name minikube plugins bridge cni hairpinmode true ipmasq true ipam ranges gateway subnet routes dst type host local isgateway true type bridge capabilities portmappings true type portmap backend type firewall type tuning capabilities aliases true domainname dns podman type dnsname stdout cli runner go run sudo n podman system info format json info go podman info host buildahversion cgroupversion conmon package conmon path usr bin conmon version conmon version commit distribution distribution fedora version memfree memtotal ociruntime name crun package crun path usr bin crun version crun version commit spec systemd selinux apparmor cap seccomp ebpf criu yajl swapfree swaptotal arch cpus eventlogger journald hostname zdalisv dfw ibm com kernel os linux rootless false uptime approximately days registries search store configfile etc containers storage conf containerstore number graphdrivername overlay graphoptions graphroot var lib containers storage graphstatus backingfilesystem extfs nativeoverlaydiff true supportsdtype true usingmetacopy false imagestore number runroot run containers storage volumepath var lib containers storage volumes cli runner go run sudo n podman container inspect f networksettings ipaddress minikube cli runner go sudo n podman container inspect f networksettings ipaddress minikube returned with exit code ssh runner go run sh c df h var awk nr print cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code retry go will retry after new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube cli runner go run sudo n podman version format version cli runner go run sudo n podman container inspect f index index networksettings ports tcp hostport minikube cli runner go sudo n podman container inspect f index index networksettings ports tcp hostport minikube returned with exit code start go error running df h var newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube start go error getting percentage of var that is free newsession new client new client error creating new ssh host from driver error getting ssh port for driver get ssh host port get port for minikube sudo n podman container inspect f index index networksettings ports tcp hostport minikube exit status stdout stderr error error inspecting object no such container minikube fix go fixhost completed within start go releasing machines lock for minikube held for out go 😿 failed to start podman container running minikube delete may fix it podman inspect ip minikube sudo n podman container inspect f networksettings ipaddress minikube exit status stdout stderr error error inspecting object no such container minikube out go out go ❌ exiting due to guest provision failed to start host podman inspect ip minikube sudo n podman container inspect f minikube exit status stdout stderr error error inspecting object no such container minikube out go out go ╭────────────────────────────────────────────────────────────────────╮ out go │ │ out go │ 😿 if the above advice does not help please let us know │ out go │ 👉 │ out go │ │ out go │ please attach the following file to the github issue │ out go │ home fedora minikube logs laststart txt │ out go │ │ out go ╰────────────────────────────────────────────────────────────────────╯ out go ❌ exiting due to guest status state unknown state minikube sudo n podman container inspect minikube format exit status stdout stderr error error inspecting object no such container minikube ╭───────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 if the above advice does not help please let us know │ │ 👉 │ │ │ │ please attach the following file to the github issue │ │ tmp minikube logs log │ │ │ ╰───────────────────────────────────────────────────────────────────────────╯ full output of minikube logs command for audit command args profile user version start time end time config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc start help minikube fedora tue jun utc tue jun utc start help minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver podman minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc config set driver minikube fedora tue jun utc tue jun utc delete minikube fedora tue jun utc tue jun utc last start log file created at running on machine binary built with gc for linux log line format mmdd hh mm ss uuuuuu threadid file line msg out go setting outfile to fd out go isatty isterminal true out go setting errfile to fd out go isatty isterminal true root go updating path home fedora minikube bin out go setting json to false start go hostinfo hostname zdalisv dfw ibm com uptime boottime procs os linux platform fedora platformfamily fedora platformversion kernelversion kernelarch virtualizationsystem kvm virtualizationrole host hostid start go virtualization kvm host out go 😄 minikube on fedora notify go checking for updates driver go setting default libvirt uri to qemu system out go ✨ using the driver based on user configuration start go selected driver start go validating driver against start go status for installed true healthy false running true needsimprovement false error libvirt group membership check failed error getting current user s gids user groupids requires cgo reason pr kvm user permission fix check that libvirtd is properly installed and that you are a member of the appropriate libvirt group remember to relogin for group changes to take effect doc out go out go 🚫 exiting due to pr kvm user permission libvirt group membership check failed error getting current user s gids user groupids requires cgo out go 💡 suggestion ensure that you are a member of the appropriate libvirt group remember to relogin for group changes to take effect out go 📘 documentation out go 🍿 related issues out go ▪ out go ▪ 🤷 profile minikube not found run minikube profile list to view all profiles 👉 to start a cluster run minikube start
0
104,757
22,749,636,943
IssuesEvent
2022-07-07 12:08:25
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
opened
Short solution needed: "golang regexp" (golang)
help wanted good first issue code golang
Please help us write most modern and shortest code solution for this issue: **golang regexp** (technology: [golang](https://onelinerhub.com/golang)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "golang regexp" (golang) - Please help us write most modern and shortest code solution for this issue: **golang regexp** (technology: [golang](https://onelinerhub.com/golang)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
non_test
short solution needed golang regexp golang please help us write most modern and shortest code solution for this issue golang regexp technology fast way just write the code solution in the comments prefered way create with a new code file inside don t forget to explain solution link to this issue in comments of pull request
0
160,887
12,520,964,395
IssuesEvent
2020-06-03 16:42:45
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739
bandit bug components software robot component: ROS static analysis testing triage version: melodic
```yaml { "id": 1, "title": "Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: melodic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-03 (16:42)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-03 (16:42)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739 - ```yaml { "id": 1, "title": "Using xml, ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called ./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: melodic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-03 (16:42)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-03 (16:42)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/ros_comm/roslaunch/src/roslaunch/xmlloader.py:739", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
test
using xml src ros comm roslaunch src roslaunch xmlloader py yaml id title using xml src ros comm roslaunch src roslaunch xmlloader py type bug description high confidence of medium severity bug using xml dom minidom parse to parse untrusted xml data is known to be vulnerable to xml attacks replace xml dom minidom parse with its defusedxml equivalent function or make sure defusedxml defuse stdlib is called src ros comm roslaunch src roslaunch xmlloader py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version melodic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src ros comm roslaunch src roslaunch xmlloader py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
1
201,101
15,173,861,749
IssuesEvent
2021-02-13 15:59:53
XProger/OpenLara
https://api.github.com/repos/XProger/OpenLara
closed
vag decoder crash at titlescreen (OpenBSD)
crash testing
``` Program received signal SIGABRT, Aborted. [Switching to thread 414848] thrkill () at /tmp/-:3 3 /tmp/-: No such file or directory. in /tmp/- Current language: auto; currently asm (gdb) bt #0 thrkill () at /tmp/-:3 #1 0x000003af38f85b7e in _libc_abort () at /usr/src/lib/libc/stdlib/abort.c:51 #2 0x000003af38f70945 in memcpy (dst0=Variable "dst0" is not available.) at /usr/src/lib/libc/string/memcpy.c:74 #3 0x000003acb6fbfb56 in Sound::VAG::decode (this=0x3af4756d000, frames=0x3af054f4000, count=1024) at sound.h:530 #4 0x000003acb6fc163e in Sound::Sample::render (this=0x3aeee409dc0, frames=0x3af054f4000, count=1024) at sound.h:954 #5 0x000003acb6fc1d46 in Sound::renderChannels (result=0x3af7ab1f000, count=1024, music=false) at sound.h:1064 #6 0x000003acb6fc2244 in Sound::fill (frames=0x3af56c04000, count=1024) at sound.h:1113 #7 0x000003acb7052b21 in sndFill (udata=0x0, stream=0x3aef01e1000 "\\", len=4096) at main.cpp:32 #8 0x000003aedaca7fe3 in ?? () from /usr/local/lib/libSDL2.so.0.8 #9 0x000003aedad11bf0 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8 #10 0x000003aedad95cf7 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8 #11 0x000003af1d7a3e21 in _rthread_start (v=Variable "v" is not available.) at /usr/src/lib/librthread/rthread.c:96 #12 0x000003af38fa8f78 in __tfork_thread () at /usr/src/lib/libc/arch/amd64/sys/tfork_thread.S:77 #13 0x0000000000000000 in ?? () (gdb) c Continuing. Program terminated with signal SIGABRT, Aborted. The program no longer exists. (gdb) q ``` directorytree and more at [pastebin](https://pastebin.com/s0dAqFwP) Could the `tmp/-: No such file or directory.` be the culprit? how to fix?
1.0
vag decoder crash at titlescreen (OpenBSD) - ``` Program received signal SIGABRT, Aborted. [Switching to thread 414848] thrkill () at /tmp/-:3 3 /tmp/-: No such file or directory. in /tmp/- Current language: auto; currently asm (gdb) bt #0 thrkill () at /tmp/-:3 #1 0x000003af38f85b7e in _libc_abort () at /usr/src/lib/libc/stdlib/abort.c:51 #2 0x000003af38f70945 in memcpy (dst0=Variable "dst0" is not available.) at /usr/src/lib/libc/string/memcpy.c:74 #3 0x000003acb6fbfb56 in Sound::VAG::decode (this=0x3af4756d000, frames=0x3af054f4000, count=1024) at sound.h:530 #4 0x000003acb6fc163e in Sound::Sample::render (this=0x3aeee409dc0, frames=0x3af054f4000, count=1024) at sound.h:954 #5 0x000003acb6fc1d46 in Sound::renderChannels (result=0x3af7ab1f000, count=1024, music=false) at sound.h:1064 #6 0x000003acb6fc2244 in Sound::fill (frames=0x3af56c04000, count=1024) at sound.h:1113 #7 0x000003acb7052b21 in sndFill (udata=0x0, stream=0x3aef01e1000 "\\", len=4096) at main.cpp:32 #8 0x000003aedaca7fe3 in ?? () from /usr/local/lib/libSDL2.so.0.8 #9 0x000003aedad11bf0 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8 #10 0x000003aedad95cf7 in SDL_DYNAPI_entry () from /usr/local/lib/libSDL2.so.0.8 #11 0x000003af1d7a3e21 in _rthread_start (v=Variable "v" is not available.) at /usr/src/lib/librthread/rthread.c:96 #12 0x000003af38fa8f78 in __tfork_thread () at /usr/src/lib/libc/arch/amd64/sys/tfork_thread.S:77 #13 0x0000000000000000 in ?? () (gdb) c Continuing. Program terminated with signal SIGABRT, Aborted. The program no longer exists. (gdb) q ``` directorytree and more at [pastebin](https://pastebin.com/s0dAqFwP) Could the `tmp/-: No such file or directory.` be the culprit? how to fix?
test
vag decoder crash at titlescreen openbsd program received signal sigabrt aborted thrkill at tmp tmp no such file or directory in tmp current language auto currently asm gdb bt thrkill at tmp in libc abort at usr src lib libc stdlib abort c in memcpy variable is not available at usr src lib libc string memcpy c in sound vag decode this frames count at sound h in sound sample render this frames count at sound h in sound renderchannels result count music false at sound h in sound fill frames count at sound h in sndfill udata stream len at main cpp in from usr local lib so in sdl dynapi entry from usr local lib so in sdl dynapi entry from usr local lib so in rthread start v variable v is not available at usr src lib librthread rthread c in tfork thread at usr src lib libc arch sys tfork thread s in gdb c continuing program terminated with signal sigabrt aborted the program no longer exists gdb q directorytree and more at could the tmp no such file or directory be the culprit how to fix
1
31,532
8,708,178,652
IssuesEvent
2018-12-06 10:09:36
indigo-dc/oidc-agent
https://api.github.com/repos/indigo-dc/oidc-agent
closed
Provide possibility to use base64 encoding instead of hex
Build Dependencies LOW Priority enhancement
base64 encoding is supported in lbisodium >= 1.0.14 So we cannot replace hex encoding with base64, but if it is supported by a system we might want to use it
1.0
Provide possibility to use base64 encoding instead of hex - base64 encoding is supported in lbisodium >= 1.0.14 So we cannot replace hex encoding with base64, but if it is supported by a system we might want to use it
non_test
provide possibility to use encoding instead of hex encoding is supported in lbisodium so we cannot replace hex encoding with but if it is supported by a system we might want to use it
0
803,855
29,191,873,213
IssuesEvent
2023-05-19 20:55:06
fabianabarca/ferias
https://api.github.com/repos/fabianabarca/ferias
closed
Crear nueva tabla para centros de origen
low_priority
Aunque actualmente los centros de origen (según Vavilov) solamente están como opciones en un campo, es posible que en el futuro queramos ampliar la información de cada uno, con datos como una reseña histórica y la ubicación geoespacial (un punto o un polígono) para un mapa, o una imagen. De forma que sería algo como: ```python class CenterOrigin(models.MODEL): name = description = region = models.PolygonField() def __str__(self): return self.name ``` También es posible que sea necesario utilizar la clasificación [actualizada](https://en.wikipedia.org/wiki/Vavilov_center).
1.0
Crear nueva tabla para centros de origen - Aunque actualmente los centros de origen (según Vavilov) solamente están como opciones en un campo, es posible que en el futuro queramos ampliar la información de cada uno, con datos como una reseña histórica y la ubicación geoespacial (un punto o un polígono) para un mapa, o una imagen. De forma que sería algo como: ```python class CenterOrigin(models.MODEL): name = description = region = models.PolygonField() def __str__(self): return self.name ``` También es posible que sea necesario utilizar la clasificación [actualizada](https://en.wikipedia.org/wiki/Vavilov_center).
non_test
crear nueva tabla para centros de origen aunque actualmente los centros de origen según vavilov solamente están como opciones en un campo es posible que en el futuro queramos ampliar la información de cada uno con datos como una reseña histórica y la ubicación geoespacial un punto o un polígono para un mapa o una imagen de forma que sería algo como python class centerorigin models model name description region models polygonfield def str self return self name también es posible que sea necesario utilizar la clasificación
0
40,714
6,846,066,341
IssuesEvent
2017-11-13 10:37:37
MarinaGarme/GESPRO_PracticaGestionTareas_1718
https://api.github.com/repos/MarinaGarme/GESPRO_PracticaGestionTareas_1718
closed
Documentación de crear estructura
documentation
Definir en la wiki una página con las diferentes secciones que deben contener la documentación.
1.0
Documentación de crear estructura - Definir en la wiki una página con las diferentes secciones que deben contener la documentación.
non_test
documentación de crear estructura definir en la wiki una página con las diferentes secciones que deben contener la documentación
0
126,676
17,093,492,211
IssuesEvent
2021-07-08 21:00:51
PowerShell/PowerShell
https://api.github.com/repos/PowerShell/PowerShell
closed
Allow .ps1xml files in XML 1.1
Issue-Enhancement Resolution-By Design WG-Cmdlets-Core
To enable colored output in output formats, I can write a Format.ps1xml file that uses ANSI color codes like so: ```xml <?xml version="1.0" encoding="utf-8"?> <Configuration> <ViewDefinitions> <!-- Resembles the git log output with format=medium (default) --> <View> <Name>Medium</Name> <ViewSelectedBy> <TypeName>PowerGit.CommitInfo</TypeName> </ViewSelectedBy> <ListControl> <ListEntries> <ListEntry> <ListItems> <ListItem> <!-- Yellow --> <Label>&#27;[33mSha</Label> <ScriptBlock>"$($_.Sha)$([char]0x001b)[0m"</ScriptBlock> </ListItem> <ListItem> <Label>Author</Label> <PropertyName>Author</PropertyName> </ListItem> <ListItem> <Label>Date</Label> <ScriptBlock>"$($_.Author.When)"</ScriptBlock> </ListItem> <ListItem> <Label>Message</Label> <PropertyName>Message</PropertyName> </ListItem> </ListItems> </ListEntry> </ListEntries> </ListControl> </View> </ViewDefinitions> </Configuration> ``` Note the `&#27;[33m`, which is `ESC[33m`, which is the ANSI code for yellow, and `$([char]0x001b)[0m`, which is `ESC[0m`, which is the ANSI reset code. Together, this will color the whole line yellow: <img width="452" alt="image" src="https://user-images.githubusercontent.com/10532611/44112434-6167e756-a005-11e8-8c14-895d27beba56.png"> just like in `git log` output: <img width="345" alt="image" src="https://user-images.githubusercontent.com/10532611/44112448-6b503818-a005-11e8-88da-5e6dabca365d.png"> HOWEVER this is actually not valid XML. Every good XML parser will rightfully complain: <img width="378" alt="image" src="https://user-images.githubusercontent.com/10532611/44112545-9eedc992-a005-11e8-8aab-0b435ab94fa1.png"> See https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.0 Which means these format files cannot be statically checked, generated, codemodded, etc. It _is_ valid in XML 1.1: https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.1 But if I change the parsing instruction to ```xml <?xml version="1.1" encoding="utf-8"?> ``` PowerShell errors: ```powershell Import-Module : Errors occurred while loading the format data file: /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml, Error in file /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml: Version number '1.1' is invalid. Line 1, position 16. At /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Import-PowerGit.ps1:32 char:1 + Import-Module -Name (Join-Path -Path $PSScriptRoot -ChildPath 'PowerG ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : InvalidOperation: (:) [Import-Module], RuntimeException + FullyQualifiedErrorId : FormatXmlUpdateException,Microsoft.PowerShell.Commands.ImportModuleCommand ``` It effectively already parses XML 1.1 fine though, it just complains about the version.
1.0
Allow .ps1xml files in XML 1.1 - To enable colored output in output formats, I can write a Format.ps1xml file that uses ANSI color codes like so: ```xml <?xml version="1.0" encoding="utf-8"?> <Configuration> <ViewDefinitions> <!-- Resembles the git log output with format=medium (default) --> <View> <Name>Medium</Name> <ViewSelectedBy> <TypeName>PowerGit.CommitInfo</TypeName> </ViewSelectedBy> <ListControl> <ListEntries> <ListEntry> <ListItems> <ListItem> <!-- Yellow --> <Label>&#27;[33mSha</Label> <ScriptBlock>"$($_.Sha)$([char]0x001b)[0m"</ScriptBlock> </ListItem> <ListItem> <Label>Author</Label> <PropertyName>Author</PropertyName> </ListItem> <ListItem> <Label>Date</Label> <ScriptBlock>"$($_.Author.When)"</ScriptBlock> </ListItem> <ListItem> <Label>Message</Label> <PropertyName>Message</PropertyName> </ListItem> </ListItems> </ListEntry> </ListEntries> </ListControl> </View> </ViewDefinitions> </Configuration> ``` Note the `&#27;[33m`, which is `ESC[33m`, which is the ANSI code for yellow, and `$([char]0x001b)[0m`, which is `ESC[0m`, which is the ANSI reset code. Together, this will color the whole line yellow: <img width="452" alt="image" src="https://user-images.githubusercontent.com/10532611/44112434-6167e756-a005-11e8-8c14-895d27beba56.png"> just like in `git log` output: <img width="345" alt="image" src="https://user-images.githubusercontent.com/10532611/44112448-6b503818-a005-11e8-88da-5e6dabca365d.png"> HOWEVER this is actually not valid XML. Every good XML parser will rightfully complain: <img width="378" alt="image" src="https://user-images.githubusercontent.com/10532611/44112545-9eedc992-a005-11e8-8aab-0b435ab94fa1.png"> See https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.0 Which means these format files cannot be statically checked, generated, codemodded, etc. It _is_ valid in XML 1.1: https://en.wikipedia.org/wiki/Valid_characters_in_XML#XML_1.1 But if I change the parsing instruction to ```xml <?xml version="1.1" encoding="utf-8"?> ``` PowerShell errors: ```powershell Import-Module : Errors occurred while loading the format data file: /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml, Error in file /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Formats/PowerGit.CommitInfo.formats.ps1xml: Version number '1.1' is invalid. Line 1, position 16. At /Users/felix/src/github.com/felixfbecker/PowerGit/PowerGit/Import-PowerGit.ps1:32 char:1 + Import-Module -Name (Join-Path -Path $PSScriptRoot -ChildPath 'PowerG ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : InvalidOperation: (:) [Import-Module], RuntimeException + FullyQualifiedErrorId : FormatXmlUpdateException,Microsoft.PowerShell.Commands.ImportModuleCommand ``` It effectively already parses XML 1.1 fine though, it just complains about the version.
non_test
allow files in xml to enable colored output in output formats i can write a format file that uses ansi color codes like so xml medium powergit commitinfo sha author author date author when message message note the which is esc which is the ansi reset code together this will color the whole line yellow img width alt image src just like in git log output img width alt image src however this is actually not valid xml every good xml parser will rightfully complain img width alt image src see which means these format files cannot be statically checked generated codemodded etc it is valid in xml but if i change the parsing instruction to xml powershell errors powershell import module errors occurred while loading the format data file users felix src github com felixfbecker powergit powergit formats powergit commitinfo formats error in file users felix src github com felixfbecker powergit powergit formats powergit commitinfo formats version number is invalid line position at users felix src github com felixfbecker powergit powergit import powergit char import module name join path path psscriptroot childpath powerg categoryinfo invalidoperation runtimeexception fullyqualifiederrorid formatxmlupdateexception microsoft powershell commands importmodulecommand it effectively already parses xml fine though it just complains about the version
0
110,607
9,462,290,017
IssuesEvent
2019-04-17 15:08:45
LIBCAS/INDIHU-Exhibition
https://api.github.com/repos/LIBCAS/INDIHU-Exhibition
closed
Náhled výstavy
waiting for test
Opět nám nefunguje (na několika počítačích i v různých prohlížečích) náhled výstavy. Snad podobný problém již řešen #45 a #5 . Jedná se o výstavu: http://inqooltest.libj.cas.cz/expo/ac131145-410d-4519-a36b-1ffac4925b3c/structure
1.0
Náhled výstavy - Opět nám nefunguje (na několika počítačích i v různých prohlížečích) náhled výstavy. Snad podobný problém již řešen #45 a #5 . Jedná se o výstavu: http://inqooltest.libj.cas.cz/expo/ac131145-410d-4519-a36b-1ffac4925b3c/structure
test
náhled výstavy opět nám nefunguje na několika počítačích i v různých prohlížečích náhled výstavy snad podobný problém již řešen a jedná se o výstavu
1
148,576
11,858,139,075
IssuesEvent
2020-03-25 10:52:05
clarity-h2020/csis-technical-validation
https://api.github.com/repos/clarity-h2020/csis-technical-validation
closed
Setup cypress.io UI integration Tests
testing
Setup cypress.io UI Tests for CSIS - [x] ~~[Fix Jenkins issues](https://github.com/clarity-h2020/csis/issues/8#issuecomment-575546178)~~ - [x] ~~create new GitHub Jenkins users~~ - [x] create 1st simple UI tests (Jenkins Job)
1.0
Setup cypress.io UI integration Tests - Setup cypress.io UI Tests for CSIS - [x] ~~[Fix Jenkins issues](https://github.com/clarity-h2020/csis/issues/8#issuecomment-575546178)~~ - [x] ~~create new GitHub Jenkins users~~ - [x] create 1st simple UI tests (Jenkins Job)
test
setup cypress io ui integration tests setup cypress io ui tests for csis create new github jenkins users create simple ui tests jenkins job
1
105,027
16,623,626,777
IssuesEvent
2021-06-03 06:45:37
Thanraj/OpenSSL_1.0.1
https://api.github.com/repos/Thanraj/OpenSSL_1.0.1
opened
CVE-2014-3569 (Medium) detected in opensslOpenSSL_1_0_1
security vulnerability
## CVE-2014-3569 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary> <p> <p>Akamai fork of openssl master.</p> <p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ssl23_get_client_hello function in s23_srvr.c in OpenSSL 0.9.8zc, 1.0.0o, and 1.0.1j does not properly handle attempts to use unsupported protocols, which allows remote attackers to cause a denial of service (NULL pointer dereference and daemon crash) via an unexpected handshake, as demonstrated by an SSLv3 handshake to a no-ssl3 application with certain error handling. NOTE: this issue became relevant after the CVE-2014-3568 fix. <p>Publish Date: 2014-12-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3569>CVE-2014-3569</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569</a></p> <p>Release Date: 2014-12-24</p> <p>Fix Resolution: OpenSSL_1_0_2a</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2014-3569 (Medium) detected in opensslOpenSSL_1_0_1 - ## CVE-2014-3569 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary> <p> <p>Akamai fork of openssl master.</p> <p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/ssl/s23_srvr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ssl23_get_client_hello function in s23_srvr.c in OpenSSL 0.9.8zc, 1.0.0o, and 1.0.1j does not properly handle attempts to use unsupported protocols, which allows remote attackers to cause a denial of service (NULL pointer dereference and daemon crash) via an unexpected handshake, as demonstrated by an SSLv3 handshake to a no-ssl3 application with certain error handling. NOTE: this issue became relevant after the CVE-2014-3568 fix. <p>Publish Date: 2014-12-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3569>CVE-2014-3569</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3569</a></p> <p>Release Date: 2014-12-24</p> <p>Fix Resolution: OpenSSL_1_0_2a</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in opensslopenssl cve medium severity vulnerability vulnerable library opensslopenssl akamai fork of openssl master library home page a href found in head commit a href found in base branch master vulnerable source files openssl ssl srvr c openssl ssl srvr c openssl ssl srvr c vulnerability details the get client hello function in srvr c in openssl and does not properly handle attempts to use unsupported protocols which allows remote attackers to cause a denial of service null pointer dereference and daemon crash via an unexpected handshake as demonstrated by an handshake to a no application with certain error handling note this issue became relevant after the cve fix publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution openssl step up your open source security game with whitesource
0
81,523
10,240,698,391
IssuesEvent
2019-08-19 21:31:15
ProjectEvergreen/greenwood
https://api.github.com/repos/ProjectEvergreen/greenwood
opened
404 page not building and serializing correctly
bug documentation question
## Type of Change - [ ] New Feature Request - [ ] Documentation / Website - [ ] Improvement / Suggestion - [x] Bug - [ ] Other (please clarify below) ## Summary Observed a couple issues with 404 page, the default anyway. Observed from [my own project](https://github.com/thegreenhouseio/www.thegreenhouse.io/issues/112) but also in getting started companion repo. Probably exists in _www/_ too. **develop** 1. `<title>` is default value, _My App_. Not sure if there could be more missing config items, like `meta` or others? <img width="310" alt="Screen Shot 2019-08-19 at 5 29 38 PM" src="https://user-images.githubusercontent.com/895923/63300781-09211200-c2a7-11e9-8efb-bae93e65c0af.png"> **build** 1. It doesn't seem to exist at all? <img width="703" alt="Screen Shot 2019-08-19 at 5 29 33 PM" src="https://user-images.githubusercontent.com/895923/63300768-032b3100-c2a7-11e9-90d5-c780fe0826d7.png"> ## Details Maybe this is more of a need around documentation as well? Not 100% clear what the expectations are. Do users make their own 404.md?
1.0
404 page not building and serializing correctly - ## Type of Change - [ ] New Feature Request - [ ] Documentation / Website - [ ] Improvement / Suggestion - [x] Bug - [ ] Other (please clarify below) ## Summary Observed a couple issues with 404 page, the default anyway. Observed from [my own project](https://github.com/thegreenhouseio/www.thegreenhouse.io/issues/112) but also in getting started companion repo. Probably exists in _www/_ too. **develop** 1. `<title>` is default value, _My App_. Not sure if there could be more missing config items, like `meta` or others? <img width="310" alt="Screen Shot 2019-08-19 at 5 29 38 PM" src="https://user-images.githubusercontent.com/895923/63300781-09211200-c2a7-11e9-8efb-bae93e65c0af.png"> **build** 1. It doesn't seem to exist at all? <img width="703" alt="Screen Shot 2019-08-19 at 5 29 33 PM" src="https://user-images.githubusercontent.com/895923/63300768-032b3100-c2a7-11e9-90d5-c780fe0826d7.png"> ## Details Maybe this is more of a need around documentation as well? Not 100% clear what the expectations are. Do users make their own 404.md?
non_test
page not building and serializing correctly type of change new feature request documentation website improvement suggestion bug other please clarify below summary observed a couple issues with page the default anyway observed from but also in getting started companion repo probably exists in www too develop is default value my app not sure if there could be more missing config items like meta or others img width alt screen shot at pm src build it doesn t seem to exist at all img width alt screen shot at pm src details maybe this is more of a need around documentation as well not clear what the expectations are do users make their own md
0
102,983
8,872,905,759
IssuesEvent
2019-01-11 16:35:17
LiskHQ/lisk-hub
https://api.github.com/repos/LiskHQ/lisk-hub
closed
Register integration tests fail randomly
type: test
Sometimes this test fail: ``` FAILED TESTS: @integration: Register Scenario: should allow to create a second passphrase ✖ Then I should be logged in HeadlessChrome 70.0.3538 (Linux 0.0.0) AssertionError: expected activePeerSet to have been called with arguments match(passphrase: punch indicate carbon slush gospel ask museum file actress spawn merit mechanic) at Helper.checkIfRegistrationConfirmed (webpack:///test/integration/register.test.js:76:44 <- src/tests.js:403066:58) at Context.<anonymous> (webpack:///test/integration/register.test.js:133:52 <- src/tests.js:403154:21) at Context.sync (webpack:///node_modules/mocha-steps/lib/step.js:29:0 <- src/tests.js:89665:24) ``` https://jenkins.lisk.io/job/lisk-hub/job/1.5.0/23/consoleFull
1.0
Register integration tests fail randomly - Sometimes this test fail: ``` FAILED TESTS: @integration: Register Scenario: should allow to create a second passphrase ✖ Then I should be logged in HeadlessChrome 70.0.3538 (Linux 0.0.0) AssertionError: expected activePeerSet to have been called with arguments match(passphrase: punch indicate carbon slush gospel ask museum file actress spawn merit mechanic) at Helper.checkIfRegistrationConfirmed (webpack:///test/integration/register.test.js:76:44 <- src/tests.js:403066:58) at Context.<anonymous> (webpack:///test/integration/register.test.js:133:52 <- src/tests.js:403154:21) at Context.sync (webpack:///node_modules/mocha-steps/lib/step.js:29:0 <- src/tests.js:89665:24) ``` https://jenkins.lisk.io/job/lisk-hub/job/1.5.0/23/consoleFull
test
register integration tests fail randomly sometimes this test fail failed tests integration register scenario should allow to create a second passphrase ✖ then i should be logged in headlesschrome linux assertionerror expected activepeerset to have been called with arguments match passphrase punch indicate carbon slush gospel ask museum file actress spawn merit mechanic at helper checkifregistrationconfirmed webpack test integration register test js src tests js at context webpack test integration register test js src tests js at context sync webpack node modules mocha steps lib step js src tests js
1
33,604
27,616,422,801
IssuesEvent
2023-03-09 19:44:59
google/site-kit-wp
https://api.github.com/repos/google/site-kit-wp
closed
Upgrade ESLint to latest
P1 Type: Infrastructure Exp: SP
## Feature Description We currently use ESLint v7 but not the latest version. Because of this, an older version of lodash is installed. See https://github.com/google/site-kit-wp/security/dependabot/34 --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * ESLint should be upgraded to the latest 7.x version (currently 7.32.0) ## Implementation Brief * Upgrade `eslint` to `^7.32.0` ### Test Coverage * No changes ## QA Brief * This only affects development tooling, so no additional QA should be necessary. We mainly need to ensure ESLint checks run as expected after upgrading, which should be observable in CR ## Changelog entry * N/A
1.0
Upgrade ESLint to latest - ## Feature Description We currently use ESLint v7 but not the latest version. Because of this, an older version of lodash is installed. See https://github.com/google/site-kit-wp/security/dependabot/34 --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * ESLint should be upgraded to the latest 7.x version (currently 7.32.0) ## Implementation Brief * Upgrade `eslint` to `^7.32.0` ### Test Coverage * No changes ## QA Brief * This only affects development tooling, so no additional QA should be necessary. We mainly need to ensure ESLint checks run as expected after upgrading, which should be observable in CR ## Changelog entry * N/A
non_test
upgrade eslint to latest feature description we currently use eslint but not the latest version because of this an older version of lodash is installed see do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria eslint should be upgraded to the latest x version currently implementation brief upgrade eslint to test coverage no changes qa brief this only affects development tooling so no additional qa should be necessary we mainly need to ensure eslint checks run as expected after upgrading which should be observable in cr changelog entry n a
0
286,644
24,767,557,190
IssuesEvent
2022-10-22 18:25:33
wasabee-project/Wasabee-IITC
https://api.github.com/repos/wasabee-project/Wasabee-IITC
closed
Filter tasks sharing some properties and bulk setter
In Testing
Some feature requests have in common the need to color all links sharing a property, like the source portal or the assigned agent. I have some ideas for this: 1. the #318 branch to select a subset of the links in the checklist and color them 2. integrate buttons (for instance) in the link list (wrt a portal) to color all inbound or outbound links 3. create a UI to give (client-side) a color to each agent and use this data to color implicitly the rendering 4. rule base coloring (rendering won't use the `color` attribute for a link) (see https://github.com/wasabee-project/Wasabee-IITC/commit/9d9d1b672afb539e508454835545a6df30958c21 for pin color) **Properties:** - by assignment - by target - by source - by zone Related: #318 #335 #350 #191
1.0
Filter tasks sharing some properties and bulk setter - Some feature requests have in common the need to color all links sharing a property, like the source portal or the assigned agent. I have some ideas for this: 1. the #318 branch to select a subset of the links in the checklist and color them 2. integrate buttons (for instance) in the link list (wrt a portal) to color all inbound or outbound links 3. create a UI to give (client-side) a color to each agent and use this data to color implicitly the rendering 4. rule base coloring (rendering won't use the `color` attribute for a link) (see https://github.com/wasabee-project/Wasabee-IITC/commit/9d9d1b672afb539e508454835545a6df30958c21 for pin color) **Properties:** - by assignment - by target - by source - by zone Related: #318 #335 #350 #191
test
filter tasks sharing some properties and bulk setter some feature requests have in common the need to color all links sharing a property like the source portal or the assigned agent i have some ideas for this the branch to select a subset of the links in the checklist and color them integrate buttons for instance in the link list wrt a portal to color all inbound or outbound links create a ui to give client side a color to each agent and use this data to color implicitly the rendering rule base coloring rendering won t use the color attribute for a link see for pin color properties by assignment by target by source by zone related
1
286,395
8,787,624,891
IssuesEvent
2018-12-20 19:18:00
containous/traefik
https://api.github.com/repos/containous/traefik
closed
Traefik killed by OOM killer in Kubernetes
area/provider/k8s kind/bug/confirmed priority/P2
### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? - Running Traefik as Deployment (ingress controller) in K8s - Having bunch of Services targeting no Pods (thus having no related Endpoints object) - Our setup is specific in way that we downscale tens of deployments to 0 on some schedule to not waste resources of K8s cluster. ### What did you expect to see? - Traefik to work and reply with error when tried to access such a service ### What did you see instead? - Traefik leaking memory and eventually being killed by OOM killer ### Output of `traefik version`: (_What version of Traefik are you using?_) 1.7.4 ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ```toml defaultEntryPoints = ["http","https"] [entryPoints] [entryPoints.http] address = ":80" compress = true [entryPoints.http.proxyProtocol] trustedIPs = ["xxxx"] [entryPoints.https] address = ":443" compress = true [entryPoints.https.proxyProtocol] trustedIPs = ["xxxx"] [entryPoints.https.tls] minVersion = "VersionTLS11" cipherSuites = [ "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256", "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305", "TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA", "TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA", "TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA", "TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA" ] [respondingTimeouts] idleTimeout = "900s" [lifeCycle] requestAcceptGraceTimeout = "10s" graceTimeOut = "10s" [api] [rest] [kubernetes] [kubernetes.ingressEndpoint] hostname = "xxxx.xxxx" [metrics] [metrics.prometheus] [accessLog] ``` ### Logs ``` time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" ``` Tons of those ^ I suspect it being a memory leak in watcher watching the Services/Endpoints as this memory growth is specific just in this case, on our other clusters where we don't downscale Deployments like this Traefik works just fine.
1.0
Traefik killed by OOM killer in Kubernetes - ### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? - Running Traefik as Deployment (ingress controller) in K8s - Having bunch of Services targeting no Pods (thus having no related Endpoints object) - Our setup is specific in way that we downscale tens of deployments to 0 on some schedule to not waste resources of K8s cluster. ### What did you expect to see? - Traefik to work and reply with error when tried to access such a service ### What did you see instead? - Traefik leaking memory and eventually being killed by OOM killer ### Output of `traefik version`: (_What version of Traefik are you using?_) 1.7.4 ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ```toml defaultEntryPoints = ["http","https"] [entryPoints] [entryPoints.http] address = ":80" compress = true [entryPoints.http.proxyProtocol] trustedIPs = ["xxxx"] [entryPoints.https] address = ":443" compress = true [entryPoints.https.proxyProtocol] trustedIPs = ["xxxx"] [entryPoints.https.tls] minVersion = "VersionTLS11" cipherSuites = [ "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256", "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305", "TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA", "TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA", "TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA", "TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA" ] [respondingTimeouts] idleTimeout = "900s" [lifeCycle] requestAcceptGraceTimeout = "10s" graceTimeOut = "10s" [api] [rest] [kubernetes] [kubernetes.ingressEndpoint] hostname = "xxxx.xxxx" [metrics] [metrics.prometheus] [accessLog] ``` ### Logs ``` time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" time="2018-11-28T11:44:33Z" level=warning msg="Endpoints not available for xxx" ``` Tons of those ^ I suspect it being a memory leak in watcher watching the Services/Endpoints as this memory growth is specific just in this case, on our other clusters where we don't downscale Deployments like this Traefik works just fine.
non_test
traefik killed by oom killer in kubernetes do you want to request a feature or report a bug bug what did you do running traefik as deployment ingress controller in having bunch of services targeting no pods thus having no related endpoints object our setup is specific in way that we downscale tens of deployments to on some schedule to not waste resources of cluster what did you expect to see traefik to work and reply with error when tried to access such a service what did you see instead traefik leaking memory and eventually being killed by oom killer output of traefik version what version of traefik are you using what is your environment configuration arguments toml provider platform toml defaultentrypoints address compress true trustedips address compress true trustedips minversion ciphersuites tls ecdhe rsa with aes gcm tls ecdhe rsa with tls ecdhe rsa with aes cbc sha tls ecdhe rsa with aes cbc sha tls ecdhe ecdsa with aes cbc sha tls ecdhe ecdsa with aes cbc sha idletimeout requestacceptgracetimeout gracetimeout hostname xxxx xxxx logs time level warning msg endpoints not available for xxx time level warning msg endpoints not available for xxx time level warning msg endpoints not available for xxx tons of those i suspect it being a memory leak in watcher watching the services endpoints as this memory growth is specific just in this case on our other clusters where we don t downscale deployments like this traefik works just fine
0
197,196
14,911,804,112
IssuesEvent
2021-01-22 11:39:11
DanielMurphy22/SmokeTests
https://api.github.com/repos/DanielMurphy22/SmokeTests
closed
Ubuntu Plotting Smoke Tests
Linux Only Manual Tests
Before testing: - Check this testing issue relates to the OS you will test on. - If unassigned, please assign yourself as for a normal Github issue. - Please run these tests on the release package of Mantid; **not a locally built version**. Afterwards: - Comment below with any issues you came across. - If no issues were found, or they are now all resolved, please close the testing issue. - Check the master issue for this OS for other unassigned smoke tests. If you have any questions please contact the creator of this issue. :chart_with_upwards_trend: :bar_chart: * 1D plotting:[instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#displaying-1d-data) - [ ] Simple plot - [ ] Another way to plot - [ ] Adding curves to existing plots - [ ] Also, test out [waterfall](https://docs.mantidproject.org/nightly/plotting/WaterfallPlotsHelp.html#waterfall-plots) and [tiled]( https://docs.mantidproject.org/nightly/plotting/1DPlotsHelp.html#tiled-plots) - [ ] Check Toolbar buttons * 2D plotting: [instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/04_displaying_2D_data.html#displaying-2d-data) - [ ] Plot all spectra - [ ] Change colour map - [ ] [Contour plot](https://docs.mantidproject.org/nightly/plotting/ColorfillPlotsHelp.html#contour-plots) (under 3D menu) - [ ] Check Toolbar buttons * [3D plotting](https://docs.mantidproject.org/nightly/plotting/3DPlotsHelp.html): - Load some data eg `LOQ7041` from the ISIS sample data - [ ] 3D surface - [ ] 3D wire frame - [ ] Check Toolbar buttons - [ ] ## Sliceviewer (further instructions and data [here](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html)) In particular try editing the data in a workspace while it is open in Sliceviewer! ([step 10](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html#alter-the-underlying-workspace))
1.0
Ubuntu Plotting Smoke Tests - Before testing: - Check this testing issue relates to the OS you will test on. - If unassigned, please assign yourself as for a normal Github issue. - Please run these tests on the release package of Mantid; **not a locally built version**. Afterwards: - Comment below with any issues you came across. - If no issues were found, or they are now all resolved, please close the testing issue. - Check the master issue for this OS for other unassigned smoke tests. If you have any questions please contact the creator of this issue. :chart_with_upwards_trend: :bar_chart: * 1D plotting:[instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#displaying-1d-data) - [ ] Simple plot - [ ] Another way to plot - [ ] Adding curves to existing plots - [ ] Also, test out [waterfall](https://docs.mantidproject.org/nightly/plotting/WaterfallPlotsHelp.html#waterfall-plots) and [tiled]( https://docs.mantidproject.org/nightly/plotting/1DPlotsHelp.html#tiled-plots) - [ ] Check Toolbar buttons * 2D plotting: [instructions](http://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/04_displaying_2D_data.html#displaying-2d-data) - [ ] Plot all spectra - [ ] Change colour map - [ ] [Contour plot](https://docs.mantidproject.org/nightly/plotting/ColorfillPlotsHelp.html#contour-plots) (under 3D menu) - [ ] Check Toolbar buttons * [3D plotting](https://docs.mantidproject.org/nightly/plotting/3DPlotsHelp.html): - Load some data eg `LOQ7041` from the ISIS sample data - [ ] 3D surface - [ ] 3D wire frame - [ ] Check Toolbar buttons - [ ] ## Sliceviewer (further instructions and data [here](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html)) In particular try editing the data in a workspace while it is open in Sliceviewer! ([step 10](https://developer.mantidproject.org/Testing/SliceViewer/SliceViewer.html#alter-the-underlying-workspace))
test
ubuntu plotting smoke tests before testing check this testing issue relates to the os you will test on if unassigned please assign yourself as for a normal github issue please run these tests on the release package of mantid not a locally built version afterwards comment below with any issues you came across if no issues were found or they are now all resolved please close the testing issue check the master issue for this os for other unassigned smoke tests if you have any questions please contact the creator of this issue chart with upwards trend bar chart plotting simple plot another way to plot adding curves to existing plots also test out and check toolbar buttons plotting plot all spectra change colour map under menu check toolbar buttons load some data eg from the isis sample data surface wire frame check toolbar buttons sliceviewer further instructions and data in particular try editing the data in a workspace while it is open in sliceviewer
1
243,860
20,593,929,781
IssuesEvent
2022-03-05 07:11:37
freelearn2000/Project1
https://api.github.com/repos/freelearn2000/Project1
opened
Automated Issue - Node App Tests/Coverage Failed
test
Someone just pushed, oh no! Here's who did it: NadaMusthafa. [Check test report here.](https://freelearn2000.github.io/Project1/nodeapp/index.html) [Check coverage report here.](https://freelearn2000.github.io/Project1/nodeapp/coverage/lcov-report/index.html) [Check lint report here.](https://freelearn2000.github.io/Project1/nodeapp/reports/lint-results.html)
1.0
Automated Issue - Node App Tests/Coverage Failed - Someone just pushed, oh no! Here's who did it: NadaMusthafa. [Check test report here.](https://freelearn2000.github.io/Project1/nodeapp/index.html) [Check coverage report here.](https://freelearn2000.github.io/Project1/nodeapp/coverage/lcov-report/index.html) [Check lint report here.](https://freelearn2000.github.io/Project1/nodeapp/reports/lint-results.html)
test
automated issue node app tests coverage failed someone just pushed oh no here s who did it nadamusthafa
1
218,462
16,761,190,022
IssuesEvent
2021-06-13 20:27:55
AlfredoCarias/Examen-Final-Analisis-de-sistemas
https://api.github.com/repos/AlfredoCarias/Examen-Final-Analisis-de-sistemas
closed
Realizacion del documento inicial del sistema
documentation
Se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo.
1.0
Realizacion del documento inicial del sistema - Se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo.
non_test
realizacion del documento inicial del sistema se realizara el documento del sistema en donde tendra la informacion del funcionamiento del mismo
0
31,410
4,705,774,368
IssuesEvent
2016-10-13 15:21:11
PulpQE/pulp-smash
https://api.github.com/repos/PulpQE/pulp-smash
opened
Test full sync with different scenarios-
Issue Type: Discussion Issue Type: Test Case Status: Needs Work
https://pulp.plan.io/issues/1983 The sync should be full in case one of these answers are true: 1. Did the config change in any way since the last sync? - A new field should be added to the Importer model 'last_updated'. As a consequence a migration should be written. 2. Has content been removed since the last sync? 3. If there was something specified in the override_config( because it changes current sync). - A new field should be added to the Importer model 'last_override_config'. As a consequence a migration should be written. 4. Force-full option was specified.
1.0
Test full sync with different scenarios- - https://pulp.plan.io/issues/1983 The sync should be full in case one of these answers are true: 1. Did the config change in any way since the last sync? - A new field should be added to the Importer model 'last_updated'. As a consequence a migration should be written. 2. Has content been removed since the last sync? 3. If there was something specified in the override_config( because it changes current sync). - A new field should be added to the Importer model 'last_override_config'. As a consequence a migration should be written. 4. Force-full option was specified.
test
test full sync with different scenarios the sync should be full in case one of these answers are true did the config change in any way since the last sync a new field should be added to the importer model last updated as a consequence a migration should be written has content been removed since the last sync if there was something specified in the override config because it changes current sync a new field should be added to the importer model last override config as a consequence a migration should be written force full option was specified
1
332,522
29,482,290,139
IssuesEvent
2023-06-02 06:59:26
mratsim/constantine
https://api.github.com/repos/mratsim/constantine
opened
Reactivate 32-bit CI
testing :passport_control:
32-bit CI is failing recently. It may be fixed whern a new Github CI Agent is out with more recent packages that avoid the ``you have held broken packages`` error. ![image](https://github.com/mratsim/constantine/assets/22738317/904faa3b-36e5-48dd-9e3c-5a4db5a8ebef) ``` Run sudo dpkg --add-architecture i386 sudo dpkg --add-architecture i386 sudo apt-fast update -qq # Try to fix "E: Unable to correct problems, you have held broken packages." sudo apt-fast clean sudo DEBIAN_FRONTEND='noninteractive' apt-fast install \ --no-install-recommends -yq \ gcc-multilib g++-multilib \ libssl-dev:i386 libgmp-dev:i386 mkdir -p external/bin cat << EOF > external/bin/gcc #!/bin/bash exec $(which gcc) -m32 "\$@" EOF cat << EOF > external/bin/g++ #!/bin/bash exec $(which g++) -m32 "\$@" EOF chmod 755 external/bin/{gcc,g++} echo '/home/runner/work/constantine/constantine/external/bin' >> $GITHUB_PATH shell: /usr/bin/bash -e {0} [apt-fast 06:24:01] [apt-fast 06:24:01]Working... this may take a while. E: Unable to correct problems, you have held broken packages. [apt-fast 06:24:01]Package manager quit with exit code. Error: Process completed with exit code 100. ```
1.0
Reactivate 32-bit CI - 32-bit CI is failing recently. It may be fixed whern a new Github CI Agent is out with more recent packages that avoid the ``you have held broken packages`` error. ![image](https://github.com/mratsim/constantine/assets/22738317/904faa3b-36e5-48dd-9e3c-5a4db5a8ebef) ``` Run sudo dpkg --add-architecture i386 sudo dpkg --add-architecture i386 sudo apt-fast update -qq # Try to fix "E: Unable to correct problems, you have held broken packages." sudo apt-fast clean sudo DEBIAN_FRONTEND='noninteractive' apt-fast install \ --no-install-recommends -yq \ gcc-multilib g++-multilib \ libssl-dev:i386 libgmp-dev:i386 mkdir -p external/bin cat << EOF > external/bin/gcc #!/bin/bash exec $(which gcc) -m32 "\$@" EOF cat << EOF > external/bin/g++ #!/bin/bash exec $(which g++) -m32 "\$@" EOF chmod 755 external/bin/{gcc,g++} echo '/home/runner/work/constantine/constantine/external/bin' >> $GITHUB_PATH shell: /usr/bin/bash -e {0} [apt-fast 06:24:01] [apt-fast 06:24:01]Working... this may take a while. E: Unable to correct problems, you have held broken packages. [apt-fast 06:24:01]Package manager quit with exit code. Error: Process completed with exit code 100. ```
test
reactivate bit ci bit ci is failing recently it may be fixed whern a new github ci agent is out with more recent packages that avoid the you have held broken packages error run sudo dpkg add architecture sudo dpkg add architecture sudo apt fast update qq try to fix e unable to correct problems you have held broken packages sudo apt fast clean sudo debian frontend noninteractive apt fast install no install recommends yq gcc multilib g multilib libssl dev libgmp dev mkdir p external bin cat external bin gcc bin bash exec which gcc eof cat external bin g bin bash exec which g eof chmod external bin gcc g echo home runner work constantine constantine external bin github path shell usr bin bash e working this may take a while e unable to correct problems you have held broken packages package manager quit with exit code error process completed with exit code
1
213,784
16,537,103,671
IssuesEvent
2021-05-27 13:03:10
finos/waltz
https://api.github.com/repos/finos/waltz
closed
Survey Tweaks: minor fixes
QoL fixed (test & close)
**Instances Summary** - [x] color code the status **Instances** - [x] when viewing make it look nicer - [x] editing, widgets should line up better
1.0
Survey Tweaks: minor fixes - **Instances Summary** - [x] color code the status **Instances** - [x] when viewing make it look nicer - [x] editing, widgets should line up better
test
survey tweaks minor fixes instances summary color code the status instances when viewing make it look nicer editing widgets should line up better
1
52,958
10,964,450,029
IssuesEvent
2019-11-27 22:39:09
The-Squadratics/openMRS_security_project
https://api.github.com/repos/The-Squadratics/openMRS_security_project
closed
Code Analysis for SSE - Task 0 - Create Working Document
Code Review
### Code Analysis for SSE - Task 0 - Project documentation The intent of the task is to create the working document and initial tasks for the Code Analysis for SSE assignment. ### Deliverable * Create a new markdown document in which we'll track the progress of this assignment * Create the initial task issues for this assignment
1.0
Code Analysis for SSE - Task 0 - Create Working Document - ### Code Analysis for SSE - Task 0 - Project documentation The intent of the task is to create the working document and initial tasks for the Code Analysis for SSE assignment. ### Deliverable * Create a new markdown document in which we'll track the progress of this assignment * Create the initial task issues for this assignment
non_test
code analysis for sse task create working document code analysis for sse task project documentation the intent of the task is to create the working document and initial tasks for the code analysis for sse assignment deliverable create a new markdown document in which we ll track the progress of this assignment create the initial task issues for this assignment
0
466,133
13,397,419,703
IssuesEvent
2020-09-03 11:34:10
jimbertools/timby
https://api.github.com/repos/jimbertools/timby
closed
Session close message and actual session closing is out of sync.
bug priority_medium
Sometimes when people are on spotty connections, they have issues with timby saying your session is closed, but then they still send a request and update their session before it closes. These need to be in sync so the "your session is closed" message is only sent when it's actually closed. ![image](https://user-images.githubusercontent.com/59569757/89525381-6084fe80-d7e6-11ea-8272-719cd5f8ca71.png)
1.0
Session close message and actual session closing is out of sync. - Sometimes when people are on spotty connections, they have issues with timby saying your session is closed, but then they still send a request and update their session before it closes. These need to be in sync so the "your session is closed" message is only sent when it's actually closed. ![image](https://user-images.githubusercontent.com/59569757/89525381-6084fe80-d7e6-11ea-8272-719cd5f8ca71.png)
non_test
session close message and actual session closing is out of sync sometimes when people are on spotty connections they have issues with timby saying your session is closed but then they still send a request and update their session before it closes these need to be in sync so the your session is closed message is only sent when it s actually closed
0
166,683
20,719,828,321
IssuesEvent
2022-03-13 07:47:30
mheob/itsb-web
https://api.github.com/repos/mheob/itsb-web
opened
prettier-stylelint-0.4.2.tgz: 3 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prettier-stylelint-0.4.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-28469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-2.0.0.tgz | Transitive | N/A | &#10060; | | [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | N/A | &#10060; | | [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | multiple | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28469</summary> ### Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - micromatch-2.3.11.tgz - parse-glob-3.0.4.tgz - glob-base-0.3.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33623</summary> ### Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/trim-newlines/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-23382</summary> ### Vulnerable Libraries - <b>postcss-6.0.23.tgz</b>, <b>postcss-5.2.18.tgz</b></p> <p> ### <b>postcss-6.0.23.tgz</b></p> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - :x: **postcss-6.0.23.tgz** (Vulnerable Library) ### <b>postcss-5.2.18.tgz</b></p> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - postcss-less-1.1.5.tgz - :x: **postcss-5.2.18.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). <p>Publish Date: 2021-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p> <p>Release Date: 2021-04-26</p> <p>Fix Resolution: postcss - 8.2.13</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"2.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;micromatch:2.3.11;parse-glob:3.0.4;glob-base:0.3.0;glob-parent:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss-less:1.1.5;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
prettier-stylelint-0.4.2.tgz: 3 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prettier-stylelint-0.4.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-28469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-2.0.0.tgz | Transitive | N/A | &#10060; | | [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | N/A | &#10060; | | [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | multiple | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28469</summary> ### Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - micromatch-2.3.11.tgz - parse-glob-3.0.4.tgz - glob-base-0.3.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33623</summary> ### Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/trim-newlines/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-23382</summary> ### Vulnerable Libraries - <b>postcss-6.0.23.tgz</b>, <b>postcss-5.2.18.tgz</b></p> <p> ### <b>postcss-6.0.23.tgz</b></p> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - :x: **postcss-6.0.23.tgz** (Vulnerable Library) ### <b>postcss-5.2.18.tgz</b></p> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/postcss-less/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - prettier-stylelint-0.4.2.tgz (Root Library) - stylelint-8.4.0.tgz - postcss-less-1.1.5.tgz - :x: **postcss-5.2.18.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mheob/itsb-web/commit/8c2365663ed4954cfb6c733614ba8886e01b2d07">8c2365663ed4954cfb6c733614ba8886e01b2d07</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). <p>Publish Date: 2021-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p> <p>Release Date: 2021-04-26</p> <p>Fix Resolution: postcss - 8.2.13</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"2.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;micromatch:2.3.11;parse-glob:3.0.4;glob-base:0.3.0;glob-parent:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"prettier-stylelint:0.4.2;stylelint:8.4.0;postcss-less:1.1.5;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_test
prettier stylelint tgz vulnerabilities highest severity is vulnerable library prettier stylelint tgz path to dependency file frontend package json path to vulnerable library frontend node modules postcss less node modules postcss package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high glob parent tgz transitive n a high trim newlines tgz transitive n a medium multiple transitive n a details cve vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules glob parent package json dependency hierarchy prettier stylelint tgz root library stylelint tgz micromatch tgz parse glob tgz glob base tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource cve vulnerable library trim newlines tgz trim newlines from the start and or end of a string library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules trim newlines package json dependency hierarchy prettier stylelint tgz root library meow tgz x trim newlines tgz vulnerable library found in head commit a href found in base branch master vulnerability details the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim newlines step up your open source security game with whitesource cve vulnerable libraries postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules postcss package json dependency hierarchy prettier stylelint tgz root library stylelint tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules postcss less node modules postcss package json dependency hierarchy prettier stylelint tgz root library stylelint tgz postcss less tgz x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss step up your open source security game with whitesource istransitivedependency true dependencytree prettier stylelint stylelint micromatch parse glob glob base glob parent isminimumfixversionavailable true minimumfixversion glob parent isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator vulnerabilityurl istransitivedependency true dependencytree prettier stylelint meow trim newlines isminimumfixversionavailable true minimumfixversion trim newlines isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method vulnerabilityurl istransitivedependency true dependencytree prettier stylelint stylelint postcss isminimumfixversionavailable true minimumfixversion postcss isbinary false packagetype javascript node js packagename postcss packageversion packagefilepaths istransitivedependency true dependencytree prettier stylelint stylelint postcss less postcss isminimumfixversionavailable true minimumfixversion postcss isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl vulnerabilityurl
0
197,502
22,595,470,453
IssuesEvent
2022-06-29 02:11:06
interserver/mailbaby-mail-api
https://api.github.com/repos/interserver/mailbaby-mail-api
closed
WS-2021-0461 (Medium) detected in swagger-ui-3.45.1.js, swagger-ui-bundle-3.45.1.js - autoclosed
security vulnerability
## WS-2021-0461 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>swagger-ui-3.45.1.js</b>, <b>swagger-ui-bundle-3.45.1.js</b></p></summary> <p> <details><summary><b>swagger-ui-3.45.1.js</b></p></summary> <p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js</a></p> <p>Path to vulnerable library: /public/doc/swagger-ui.js</p> <p> Dependency Hierarchy: - :x: **swagger-ui-3.45.1.js** (Vulnerable Library) </details> <details><summary><b>swagger-ui-bundle-3.45.1.js</b></p></summary> <p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js</a></p> <p>Path to dependency file: /public/doc/index.html</p> <p>Path to vulnerable library: /public/doc/./swagger-ui-bundle.js,/public/doc/swagger-ui-bundle.js</p> <p> Dependency Hierarchy: - :x: **swagger-ui-bundle-3.45.1.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SwaggerUI supports displaying remote OpenAPI definitions through the ?url parameter. This enables robust demonstration capabilities on sites like petstore.swagger.io, editor.swagger.io, and similar sites, where users often want to see what their OpenAPI definitions would look like rendered. However, this functionality may pose a risk for users who host their own SwaggerUI instances. In particular, including remote OpenAPI definitions opens a vector for phishing attacks by abusing the trusted names/domains of self-hosted instances. Resolution: We've made the decision to disable query parameters (#4872) by default starting with SwaggerUI version 4.1.3. Please update to this version when it becomes available (ETA: 2021 December). Users will still be able to be re-enable the options at their discretion. We'll continue to enable query parameters on the Swagger demo sites. <p>Publish Date: 2021-12-09 <p>URL: <a href=https://github.com/swagger-api/swagger-ui/commit/01a3e55960f864a0acf6a8d06e5ddaf6776a7f76>WS-2021-0461</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-qrmm-w75w-3wpx">https://github.com/advisories/GHSA-qrmm-w75w-3wpx</a></p> <p>Release Date: 2021-12-09</p> <p>Fix Resolution: swagger-ui - 4.1.3;swagger-ui-dist - 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0461 (Medium) detected in swagger-ui-3.45.1.js, swagger-ui-bundle-3.45.1.js - autoclosed - ## WS-2021-0461 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>swagger-ui-3.45.1.js</b>, <b>swagger-ui-bundle-3.45.1.js</b></p></summary> <p> <details><summary><b>swagger-ui-3.45.1.js</b></p></summary> <p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui.js</a></p> <p>Path to vulnerable library: /public/doc/swagger-ui.js</p> <p> Dependency Hierarchy: - :x: **swagger-ui-3.45.1.js** (Vulnerable Library) </details> <details><summary><b>swagger-ui-bundle-3.45.1.js</b></p></summary> <p>Swagger UI is a dependency-free collection of HTML, JavaScript, and CSS assets that dynamically generate beautiful documentation from a Swagger-compliant API</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js">https://cdnjs.cloudflare.com/ajax/libs/swagger-ui/3.45.1/swagger-ui-bundle.js</a></p> <p>Path to dependency file: /public/doc/index.html</p> <p>Path to vulnerable library: /public/doc/./swagger-ui-bundle.js,/public/doc/swagger-ui-bundle.js</p> <p> Dependency Hierarchy: - :x: **swagger-ui-bundle-3.45.1.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SwaggerUI supports displaying remote OpenAPI definitions through the ?url parameter. This enables robust demonstration capabilities on sites like petstore.swagger.io, editor.swagger.io, and similar sites, where users often want to see what their OpenAPI definitions would look like rendered. However, this functionality may pose a risk for users who host their own SwaggerUI instances. In particular, including remote OpenAPI definitions opens a vector for phishing attacks by abusing the trusted names/domains of self-hosted instances. Resolution: We've made the decision to disable query parameters (#4872) by default starting with SwaggerUI version 4.1.3. Please update to this version when it becomes available (ETA: 2021 December). Users will still be able to be re-enable the options at their discretion. We'll continue to enable query parameters on the Swagger demo sites. <p>Publish Date: 2021-12-09 <p>URL: <a href=https://github.com/swagger-api/swagger-ui/commit/01a3e55960f864a0acf6a8d06e5ddaf6776a7f76>WS-2021-0461</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-qrmm-w75w-3wpx">https://github.com/advisories/GHSA-qrmm-w75w-3wpx</a></p> <p>Release Date: 2021-12-09</p> <p>Fix Resolution: swagger-ui - 4.1.3;swagger-ui-dist - 4.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws medium detected in swagger ui js swagger ui bundle js autoclosed ws medium severity vulnerability vulnerable libraries swagger ui js swagger ui bundle js swagger ui js swagger ui is a dependency free collection of html javascript and css assets that dynamically generate beautiful documentation from a swagger compliant api library home page a href path to vulnerable library public doc swagger ui js dependency hierarchy x swagger ui js vulnerable library swagger ui bundle js swagger ui is a dependency free collection of html javascript and css assets that dynamically generate beautiful documentation from a swagger compliant api library home page a href path to dependency file public doc index html path to vulnerable library public doc swagger ui bundle js public doc swagger ui bundle js dependency hierarchy x swagger ui bundle js vulnerable library found in head commit a href found in base branch master vulnerability details swaggerui supports displaying remote openapi definitions through the url parameter this enables robust demonstration capabilities on sites like petstore swagger io editor swagger io and similar sites where users often want to see what their openapi definitions would look like rendered however this functionality may pose a risk for users who host their own swaggerui instances in particular including remote openapi definitions opens a vector for phishing attacks by abusing the trusted names domains of self hosted instances resolution we ve made the decision to disable query parameters by default starting with swaggerui version please update to this version when it becomes available eta december users will still be able to be re enable the options at their discretion we ll continue to enable query parameters on the swagger demo sites publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution swagger ui swagger ui dist step up your open source security game with mend
0
218,847
17,026,567,590
IssuesEvent
2021-07-03 16:46:24
ZinedineMess/Billed-app-FR
https://api.github.com/repos/ZinedineMess/Billed-app-FR
closed
TU composant container/Bills
tests
- [x] couvrir tous les "statements" sauf les appels au back-end firebase (ils sont signalés en commentaire dans le code) : c'est simple, il faut que [le rapport de couverture du fichier container/Bills](http://127.0.0.1:8080/coverage/lcov-report/containers/Bills.js.html) soit vert. Cela devrait permettre d'obtenir un taux de couverture aux alentours de 80% dans la colonne "statements".
1.0
TU composant container/Bills - - [x] couvrir tous les "statements" sauf les appels au back-end firebase (ils sont signalés en commentaire dans le code) : c'est simple, il faut que [le rapport de couverture du fichier container/Bills](http://127.0.0.1:8080/coverage/lcov-report/containers/Bills.js.html) soit vert. Cela devrait permettre d'obtenir un taux de couverture aux alentours de 80% dans la colonne "statements".
test
tu composant container bills couvrir tous les statements sauf les appels au back end firebase ils sont signalés en commentaire dans le code c est simple il faut que soit vert cela devrait permettre d obtenir un taux de couverture aux alentours de dans la colonne statements
1
290,797
25,096,497,782
IssuesEvent
2022-11-08 10:31:24
Tencent/spring-cloud-tencent
https://api.github.com/repos/Tencent/spring-cloud-tencent
closed
【需求认领】提升 spring-cloud-tencent-gateway-plugin 模块的单元测试覆盖率
good first issue unit test
- 目标提升到80%以上 <img width="752" alt="image" src="https://user-images.githubusercontent.com/24446200/199162123-df7bb42d-e152-4bcd-b116-ae51aceb76fc.png"> - 如何获取单侧覆盖率: 1. 项目根目录下运行`mvn test` 2. 项目根目录下用浏览器打开`target/site/jacoco/index.html`即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充 - 额外提供 需要额外提供更新后的单测覆盖率截图到 PR 下方
1.0
【需求认领】提升 spring-cloud-tencent-gateway-plugin 模块的单元测试覆盖率 - - 目标提升到80%以上 <img width="752" alt="image" src="https://user-images.githubusercontent.com/24446200/199162123-df7bb42d-e152-4bcd-b116-ae51aceb76fc.png"> - 如何获取单侧覆盖率: 1. 项目根目录下运行`mvn test` 2. 项目根目录下用浏览器打开`target/site/jacoco/index.html`即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充 - 额外提供 需要额外提供更新后的单测覆盖率截图到 PR 下方
test
【需求认领】提升 spring cloud tencent gateway plugin 模块的单元测试覆盖率 以上 img width alt image src 如何获取单侧覆盖率: 项目根目录下运行 mvn test 项目根目录下用浏览器打开 target site jacoco index html 即可查看单侧覆盖率,点进去即可查看详细覆盖情况以作出补充 额外提供 需要额外提供更新后的单测覆盖率截图到 pr 下方
1
252,464
21,579,491,148
IssuesEvent
2022-05-02 17:06:58
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Editing Secret object in Rancher UI corrupting binary data
kind/bug [zube]: To Test internal priority/1 team/area1
Internal reference: SURE-4281 Reported in 2.6.3 The binary data in the Secret object getting corrupted after editing in the Rancher UI. The issue we have come across now is that if we use the webui to update an existing secret that contains a key/value pair with binary data and for example add a new key/value pair and save that the binary data becomes malformed. Repro steps: - Use a Secret manifest file. Create the Secret object with kubectl - go to Rancher UI and edit the secret. I added some test key/value in the UI - `kubectl get secret secret-name -o yaml` is shows different values for binary data. Before & after show control characters appearing via some conversion. Before: ``` apiVersion: v1 data: amq-client.ks: MIIURQIBAzCCE/4GCSqGSIb3DQEHAaCCE+8EghPrMIIT5zCCCsMGCSqGSIb3DQEHAaCCCrQEggqwMIIKrDCCBU4GCyqGSIb3DQEMCgECoIIE+zCCBPcwKQYKKoZIhvcNAQwBAzAbBBQvb+FEMTTSYRXDivpl53xIWp ``` After: ``` apiVersion: v1 data: amq-client.ks: MO+/vRRFAgEDMO+/vRPvv70GCSrvv71I77+977+9DQEHAe+/ve+/vRPvv70E77+9E++/vTDvv70T77+9MO+/vQrvv70GCSrvv71I77+977+9DQEHAe+/ve+/vQrvv70E77+9Cu+/vTDvv70K77+9MO+/vQVOBgsq77+9SO+/ve+/vQ0BDAoBAu+/ve+/vQTvv70w77+9BO+/vTApBgoq77+9SO+/ve+/ ``` Workaround: Is workaround available and implemented? yes What is the workaround: using the `kubectl` command to edit the secrets. Original work (for 2.4.6): https://github.com/rancher/rancher/issues/27439
1.0
Editing Secret object in Rancher UI corrupting binary data - Internal reference: SURE-4281 Reported in 2.6.3 The binary data in the Secret object getting corrupted after editing in the Rancher UI. The issue we have come across now is that if we use the webui to update an existing secret that contains a key/value pair with binary data and for example add a new key/value pair and save that the binary data becomes malformed. Repro steps: - Use a Secret manifest file. Create the Secret object with kubectl - go to Rancher UI and edit the secret. I added some test key/value in the UI - `kubectl get secret secret-name -o yaml` is shows different values for binary data. Before & after show control characters appearing via some conversion. Before: ``` apiVersion: v1 data: amq-client.ks: MIIURQIBAzCCE/4GCSqGSIb3DQEHAaCCE+8EghPrMIIT5zCCCsMGCSqGSIb3DQEHAaCCCrQEggqwMIIKrDCCBU4GCyqGSIb3DQEMCgECoIIE+zCCBPcwKQYKKoZIhvcNAQwBAzAbBBQvb+FEMTTSYRXDivpl53xIWp ``` After: ``` apiVersion: v1 data: amq-client.ks: MO+/vRRFAgEDMO+/vRPvv70GCSrvv71I77+977+9DQEHAe+/ve+/vRPvv70E77+9E++/vTDvv70T77+9MO+/vQrvv70GCSrvv71I77+977+9DQEHAe+/ve+/vQrvv70E77+9Cu+/vTDvv70K77+9MO+/vQVOBgsq77+9SO+/ve+/vQ0BDAoBAu+/ve+/vQTvv70w77+9BO+/vTApBgoq77+9SO+/ve+/ ``` Workaround: Is workaround available and implemented? yes What is the workaround: using the `kubectl` command to edit the secrets. Original work (for 2.4.6): https://github.com/rancher/rancher/issues/27439
test
editing secret object in rancher ui corrupting binary data internal reference sure reported in the binary data in the secret object getting corrupted after editing in the rancher ui the issue we have come across now is that if we use the webui to update an existing secret that contains a key value pair with binary data and for example add a new key value pair and save that the binary data becomes malformed repro steps use a secret manifest file create the secret object with kubectl go to rancher ui and edit the secret i added some test key value in the ui kubectl get secret secret name o yaml is shows different values for binary data before after show control characters appearing via some conversion before apiversion data amq client ks miiurqibazcce zccbpcwkqykkozihvcnaqwbazabbbqvb after apiversion data amq client ks mo vrrfagedmo ve ve ve ve ve workaround is workaround available and implemented yes what is the workaround using the kubectl command to edit the secrets original work for
1
287,975
31,856,865,337
IssuesEvent
2023-09-15 08:07:09
nidhi7598/linux-4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564
closed
CVE-2023-1390 (High) detected in linuxlinux-4.19.294 - autoclosed
Mend: dependency security vulnerability
## CVE-2023-1390 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A remote denial of service vulnerability was found in the Linux kernel’s TIPC kernel module. The while loop in tipc_link_xmit() hits an unknown state while attempting to parse SKBs, which are not in the queue. Sending two small UDP packets to a system with a UDP bearer results in the CPU utilization for the system to instantly spike to 100%, causing a denial of service condition. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1390>CVE-2023-1390</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1390">https://www.linuxkernelcves.com/cves/CVE-2023-1390</a></p> <p>Release Date: 2023-03-16</p> <p>Fix Resolution: v4.9.253,v4.14.217,v4.19.170,v5.4.92,v5.10.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-1390 (High) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2023-1390 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A remote denial of service vulnerability was found in the Linux kernel’s TIPC kernel module. The while loop in tipc_link_xmit() hits an unknown state while attempting to parse SKBs, which are not in the queue. Sending two small UDP packets to a system with a UDP bearer results in the CPU utilization for the system to instantly spike to 100%, causing a denial of service condition. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1390>CVE-2023-1390</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1390">https://www.linuxkernelcves.com/cves/CVE-2023-1390</a></p> <p>Release Date: 2023-03-16</p> <p>Fix Resolution: v4.9.253,v4.14.217,v4.19.170,v5.4.92,v5.10.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a remote denial of service vulnerability was found in the linux kernel’s tipc kernel module the while loop in tipc link xmit hits an unknown state while attempting to parse skbs which are not in the queue sending two small udp packets to a system with a udp bearer results in the cpu utilization for the system to instantly spike to causing a denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
239,186
19,827,802,176
IssuesEvent
2022-01-20 08:50:30
CoffeeKumazaki/arXiv
https://api.github.com/repos/CoffeeKumazaki/arXiv
closed
AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED)
Simulation Testing
AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED)<br> https://ift.tt/3nUFyja<br> <p>As self-driving systems become better, simulating scenarios where the autonomy stack may fail becomes more important. Traditionally, those scenarios are generated for a few scenes with respect to the planning module that takes ground-truth actor states as input. This does not scale and cannot identify all possible autonomy failures, such as perception failures due to occlusion. In this paper, we propose AdvSim, an adversarial framework to generate safety-critical scenarios for any LiDAR-based autonomy system. Given an initial traffic scenario, AdvSim modifies the actors' trajectories in a physically plausible manner and updates the LiDAR sensor data to match the perturbed world. Importantly, by simulating directly from sensor data, we obtain adversarial scenarios that are safety-critical for the full autonomy stack. Our experiments show that our approach is general and can identify thousands of semantically meaningful safety-critical scenarios for a wide range of modern self-driving systems. Furthermore, we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by AdvSim.</p> <br> <br> via cs.RO updates on arXiv.org http://arxiv.org/<br>
1.0
AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED) - AdvSim: Generating Safety-Critical Scenarios for Self-Driving Vehicles. (arXiv:2101.06549v3 [cs.RO] UPDATED)<br> https://ift.tt/3nUFyja<br> <p>As self-driving systems become better, simulating scenarios where the autonomy stack may fail becomes more important. Traditionally, those scenarios are generated for a few scenes with respect to the planning module that takes ground-truth actor states as input. This does not scale and cannot identify all possible autonomy failures, such as perception failures due to occlusion. In this paper, we propose AdvSim, an adversarial framework to generate safety-critical scenarios for any LiDAR-based autonomy system. Given an initial traffic scenario, AdvSim modifies the actors' trajectories in a physically plausible manner and updates the LiDAR sensor data to match the perturbed world. Importantly, by simulating directly from sensor data, we obtain adversarial scenarios that are safety-critical for the full autonomy stack. Our experiments show that our approach is general and can identify thousands of semantically meaningful safety-critical scenarios for a wide range of modern self-driving systems. Furthermore, we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by AdvSim.</p> <br> <br> via cs.RO updates on arXiv.org http://arxiv.org/<br>
test
advsim generating safety critical scenarios for self driving vehicles arxiv updated advsim generating safety critical scenarios for self driving vehicles arxiv updated as self driving systems become better simulating scenarios where the autonomy stack may fail becomes more important traditionally those scenarios are generated for a few scenes with respect to the planning module that takes ground truth actor states as input this does not scale and cannot identify all possible autonomy failures such as perception failures due to occlusion in this paper we propose advsim an adversarial framework to generate safety critical scenarios for any lidar based autonomy system given an initial traffic scenario advsim modifies the actors trajectories in a physically plausible manner and updates the lidar sensor data to match the perturbed world importantly by simulating directly from sensor data we obtain adversarial scenarios that are safety critical for the full autonomy stack our experiments show that our approach is general and can identify thousands of semantically meaningful safety critical scenarios for a wide range of modern self driving systems furthermore we show that the robustness and safety of these systems can be further improved by training them with scenarios generated by advsim via cs ro updates on arxiv org
1