Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
228,333 | 18,170,847,054 | IssuesEvent | 2021-09-27 19:50:42 | dbrownukk/EFD_v2 | https://api.github.com/repos/dbrownukk/EFD_v2 | closed | OIHM: Ensure consistency between drop downs in household interview data entry spreadsheets and direct entry into the web app | enhancement For Testing BacklogCandidate | QA instance
Household data entry OHEA and OIHM
Assets
Land
There as a ? in the Unit box . This may cause confusion.
Could a drop down be added with Acre or Hectare as the options
| 1.0 | OIHM: Ensure consistency between drop downs in household interview data entry spreadsheets and direct entry into the web app - QA instance
Household data entry OHEA and OIHM
Assets
Land
There as a ? in the Unit box . This may cause confusion.
Could a drop down be added with Acre or Hectare as the options
| test | oihm ensure consistency between drop downs in household interview data entry spreadsheets and direct entry into the web app qa instance household data entry ohea and oihm assets land there as a in the unit box this may cause confusion could a drop down be added with acre or hectare as the options | 1 |
184,347 | 21,784,873,068 | IssuesEvent | 2022-05-14 01:38:02 | onokatio/blog.katio.net | https://api.github.com/repos/onokatio/blog.katio.net | closed | WS-2021-0154 (Medium) detected in glob-parent-3.1.0.tgz - autoclosed | security vulnerability | ## WS-2021-0154 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: blog.katio.net/package.json</p>
<p>Path to vulnerable library: blog.katio.net/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- cli-7.10.5.tgz (Root Library)
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) vulnerability was found in glob-parent before 5.1.2.
<p>Publish Date: 2021-01-27
<p>URL: <a href=https://github.com/gulpjs/glob-parent/commit/f9231168b0041fea3f8f954b3cceb56269fc6366>WS-2021-0154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/releases/tag/v5.1.2">https://github.com/gulpjs/glob-parent/releases/tag/v5.1.2</a></p>
<p>Release Date: 2021-01-27</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0154 (Medium) detected in glob-parent-3.1.0.tgz - autoclosed - ## WS-2021-0154 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: blog.katio.net/package.json</p>
<p>Path to vulnerable library: blog.katio.net/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- cli-7.10.5.tgz (Root Library)
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) vulnerability was found in glob-parent before 5.1.2.
<p>Publish Date: 2021-01-27
<p>URL: <a href=https://github.com/gulpjs/glob-parent/commit/f9231168b0041fea3f8f954b3cceb56269fc6366>WS-2021-0154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/releases/tag/v5.1.2">https://github.com/gulpjs/glob-parent/releases/tag/v5.1.2</a></p>
<p>Release Date: 2021-01-27</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in glob parent tgz autoclosed ws medium severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file blog katio net package json path to vulnerable library blog katio net node modules glob parent package json dependency hierarchy cli tgz root library chokidar tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details regular expression denial of service redos vulnerability was found in glob parent before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource | 0 |
338,855 | 30,324,791,702 | IssuesEvent | 2023-07-10 22:34:20 | microsoft/vscode-python | https://api.github.com/repos/microsoft/vscode-python | closed | new pytest runner mixes test classes | bug area-testing needs PR |
Type: <b>Bug</b>
<!-- Please fill in all XXX markers -->
# Behaviour
## Expected vs. Actual
I have enabled the experiment `pythonTestAdapter` to use the new test runner. If two test files have identically named test classes, their test methods are mixed
## Steps to reproduce:
1. Create these two files:
```python
# tests/test_a.py
class TestSomething:
def test_a(self):
assert True
```
```python
# tests/test_b.py
class TestSomething:
def test_b(self):
assert True
```
2. Refresh the test suites
3. You should get the following test structure:
```
test_a.py
> TestSomething
> test_a
> test_b
test_b.py
> TestSomething
> test_a
> test_b
```
instead of the expected
```
test_a.py
> TestSomething
> test_a
test_b.py
> TestSomething
> test_b
```
<!--
**After** creating the issue on GitHub, you can add screenshots and GIFs of what is happening. Consider tools like https://www.cockos.com/licecap/, https://github.com/phw/peek or https://www.screentogif.com/ for GIF creation.
-->
<!-- **NOTE**: Everything below except Python output panel is auto-generated; no editing required. Please do provide Python output panel. -->
# Diagnostic data
- Python version (& distribution if applicable, e.g. Anaconda): 3.9.2
- Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Venv
- Value of the `python.languageServer` setting: Default
<details>
<summary>Output for <code>Python</code> in the <code>Output</code> panel (<code>View</code>→<code>Output</code>, change the drop-down the upper-right of the <code>Output</code> panel to <code>Python</code>)
</summary>
<p>
```
XXX
```
</p>
</details>
<details>
<summary>User Settings</summary>
<p>
```
Multiroot scenario, following user settings may not apply:
languageServer: "Pylance"
formatting
• provider: "yapf"
experiments
• optInto: ["pythonTestAdapter"]
```
</p>
</details>
Extension version: 2023.12.0
VS Code version: Code 1.80.0 (660393deaaa6d1996740ff4880f1bad43768c814, 2023-07-04T13:39:48.236Z)
OS version: Linux x64 5.15.0-75-generic
Modes:
Remote OS version: Linux x64 5.15.0-75-generic
<details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
vslsvsres303:30308271
vserr242cf:30382550
pythontb:30283811
vsjup518:30340749
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscod805:30301674
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593cf:30376535
pythonvs932:30410667
py29gd2263:30776702
vscaat:30438848
vsclangdc:30486549
c4g48928:30535728
dsvsc012:30540252
pynewext54:30695312
azure-dev_surveyone:30548225
vsccc:30610678
2e4cg342:30602488
pyind779:30671433
89544117:30613380
pythonsymbol12:30671437
2i9eh265:30646982
showlangstatbar:30737416
vsctsb:30748421
pythonms35:30701012
03d35959:30757346
pythonfmttext:30731395
pythoncmvfstrcf:30756944
fixshowwlkth:30771522
hideindicator:30766889
pythongtdpath:30769146
bgfeh915:30780428
pythonnosmt12:30779714
pythonidxpt:30784022
pythonnoceb:30776495
e537b577:30772215
```
</details>
<!-- generated by issue reporter --> | 1.0 | new pytest runner mixes test classes -
Type: <b>Bug</b>
<!-- Please fill in all XXX markers -->
# Behaviour
## Expected vs. Actual
I have enabled the experiment `pythonTestAdapter` to use the new test runner. If two test files have identically named test classes, their test methods are mixed
## Steps to reproduce:
1. Create these two files:
```python
# tests/test_a.py
class TestSomething:
def test_a(self):
assert True
```
```python
# tests/test_b.py
class TestSomething:
def test_b(self):
assert True
```
2. Refresh the test suites
3. You should get the following test structure:
```
test_a.py
> TestSomething
> test_a
> test_b
test_b.py
> TestSomething
> test_a
> test_b
```
instead of the expected
```
test_a.py
> TestSomething
> test_a
test_b.py
> TestSomething
> test_b
```
<!--
**After** creating the issue on GitHub, you can add screenshots and GIFs of what is happening. Consider tools like https://www.cockos.com/licecap/, https://github.com/phw/peek or https://www.screentogif.com/ for GIF creation.
-->
<!-- **NOTE**: Everything below except Python output panel is auto-generated; no editing required. Please do provide Python output panel. -->
# Diagnostic data
- Python version (& distribution if applicable, e.g. Anaconda): 3.9.2
- Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Venv
- Value of the `python.languageServer` setting: Default
<details>
<summary>Output for <code>Python</code> in the <code>Output</code> panel (<code>View</code>→<code>Output</code>, change the drop-down the upper-right of the <code>Output</code> panel to <code>Python</code>)
</summary>
<p>
```
XXX
```
</p>
</details>
<details>
<summary>User Settings</summary>
<p>
```
Multiroot scenario, following user settings may not apply:
languageServer: "Pylance"
formatting
• provider: "yapf"
experiments
• optInto: ["pythonTestAdapter"]
```
</p>
</details>
Extension version: 2023.12.0
VS Code version: Code 1.80.0 (660393deaaa6d1996740ff4880f1bad43768c814, 2023-07-04T13:39:48.236Z)
OS version: Linux x64 5.15.0-75-generic
Modes:
Remote OS version: Linux x64 5.15.0-75-generic
<details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
vslsvsres303:30308271
vserr242cf:30382550
pythontb:30283811
vsjup518:30340749
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscod805:30301674
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593cf:30376535
pythonvs932:30410667
py29gd2263:30776702
vscaat:30438848
vsclangdc:30486549
c4g48928:30535728
dsvsc012:30540252
pynewext54:30695312
azure-dev_surveyone:30548225
vsccc:30610678
2e4cg342:30602488
pyind779:30671433
89544117:30613380
pythonsymbol12:30671437
2i9eh265:30646982
showlangstatbar:30737416
vsctsb:30748421
pythonms35:30701012
03d35959:30757346
pythonfmttext:30731395
pythoncmvfstrcf:30756944
fixshowwlkth:30771522
hideindicator:30766889
pythongtdpath:30769146
bgfeh915:30780428
pythonnosmt12:30779714
pythonidxpt:30784022
pythonnoceb:30776495
e537b577:30772215
```
</details>
<!-- generated by issue reporter --> | test | new pytest runner mixes test classes type bug behaviour expected vs actual i have enabled the experiment pythontestadapter to use the new test runner if two test files have identically named test classes their test methods are mixed steps to reproduce create these two files python tests test a py class testsomething def test a self assert true python tests test b py class testsomething def test b self assert true refresh the test suites you should get the following test structure test a py testsomething test a test b test b py testsomething test a test b instead of the expected test a py testsomething test a test b py testsomething test b after creating the issue on github you can add screenshots and gifs of what is happening consider tools like or for gif creation diagnostic data python version distribution if applicable e g anaconda type of virtual environment used e g conda venv virtualenv etc venv value of the python languageserver setting default output for python in the output panel view → output change the drop down the upper right of the output panel to python xxx user settings multiroot scenario following user settings may not apply languageserver pylance formatting • provider yapf experiments • optinto extension version vs code version code os version linux generic modes remote os version linux generic a b experiments pythontb pythonptprofiler vscaat vsclangdc azure dev surveyone vsccc showlangstatbar vsctsb pythonfmttext pythoncmvfstrcf fixshowwlkth hideindicator pythongtdpath pythonidxpt pythonnoceb | 1 |
314,252 | 23,512,487,401 | IssuesEvent | 2022-08-18 17:58:28 | cal-itp/data-infra | https://api.github.com/repos/cal-itp/data-infra | closed | Docs request: what even is a transit agency and how can we answer questions that relate to their GTFS data? | type: documentation | A lot of times Caltrans staff will get asked various questions such as #984. When attempting to answer this question a lot of confusion can arise with how to quantify this answer due to our thorough dataset in the California Transit airtable. A quick how-to guide is needed to determine:
1. What organizations do we want to analyze depending on the research question?
There has been an effort (see the [Transit Provider Categorization](https://docs.google.com/document/d/1e0Z-aLKzNg7xnP0gSCCSWOuNPjTB8xN1G7vWVYQ8bYk/edit#heading=h.5taosqpsjhbq) doc) to classify organizations into various tracking, assistance or reporting categories. A shorter version of this should be documented such that analysts can quickly know what kind of filtering of organizations to use depending on the research ask.
2. How should an organization's services be aggregated when analyzing?
Typical research questions ask what seem to be fairly basic questions like #984. However, an organization can have multiple services that it manages and a service can be described in multiple GTFS datasets. Therefore, a situation can arise where it is unclear how the completeness of data should be counted. Is it when all services that an organization manages have a certain quality about them or when at least one, or a threshold? A clear plan of what aggregation strategy to use should be agreed upon and documented. | 1.0 | Docs request: what even is a transit agency and how can we answer questions that relate to their GTFS data? - A lot of times Caltrans staff will get asked various questions such as #984. When attempting to answer this question a lot of confusion can arise with how to quantify this answer due to our thorough dataset in the California Transit airtable. A quick how-to guide is needed to determine:
1. What organizations do we want to analyze depending on the research question?
There has been an effort (see the [Transit Provider Categorization](https://docs.google.com/document/d/1e0Z-aLKzNg7xnP0gSCCSWOuNPjTB8xN1G7vWVYQ8bYk/edit#heading=h.5taosqpsjhbq) doc) to classify organizations into various tracking, assistance or reporting categories. A shorter version of this should be documented such that analysts can quickly know what kind of filtering of organizations to use depending on the research ask.
2. How should an organization's services be aggregated when analyzing?
Typical research questions ask what seem to be fairly basic questions like #984. However, an organization can have multiple services that it manages and a service can be described in multiple GTFS datasets. Therefore, a situation can arise where it is unclear how the completeness of data should be counted. Is it when all services that an organization manages have a certain quality about them or when at least one, or a threshold? A clear plan of what aggregation strategy to use should be agreed upon and documented. | non_test | docs request what even is a transit agency and how can we answer questions that relate to their gtfs data a lot of times caltrans staff will get asked various questions such as when attempting to answer this question a lot of confusion can arise with how to quantify this answer due to our thorough dataset in the california transit airtable a quick how to guide is needed to determine what organizations do we want to analyze depending on the research question there has been an effort see the doc to classify organizations into various tracking assistance or reporting categories a shorter version of this should be documented such that analysts can quickly know what kind of filtering of organizations to use depending on the research ask how should an organization s services be aggregated when analyzing typical research questions ask what seem to be fairly basic questions like however an organization can have multiple services that it manages and a service can be described in multiple gtfs datasets therefore a situation can arise where it is unclear how the completeness of data should be counted is it when all services that an organization manages have a certain quality about them or when at least one or a threshold a clear plan of what aggregation strategy to use should be agreed upon and documented | 0 |
290,880 | 25,101,976,435 | IssuesEvent | 2022-11-08 14:13:52 | pingcap/tiflow | https://api.github.com/repos/pingcap/tiflow | closed | Unstable integration test e2e_with_selectors | type/bug component/test severity/minor area/engine | ### Which jobs are flaking?
dataflow engine integration test
### Which test(s) are flaking?
e2e_with_selectors
### Jenkins logs or GitHub Actions link
https://ci2.pingcap.net/blue/organizations/jenkins/engine_ghpr_integration_test/detail/engine_ghpr_integration_test/1774/pipeline
detail logs:
[log-e2e_with_selectors-dm_many_tables.tar.gz](https://github.com/pingcap/tiflow/files/9959629/log-e2e_with_selectors-dm_many_tables.tar.gz)
```go
[2022-11-08T07:52:47.555Z] create_job_json: {
[2022-11-08T07:52:47.555Z] "type": "FakeJob",
[2022-11-08T07:52:47.555Z] "selectors": [
[2022-11-08T07:52:47.555Z] {
[2022-11-08T07:52:47.555Z] "label": "name",
[2022-11-08T07:52:47.555Z] "target": "exec-1",
[2022-11-08T07:52:47.555Z] "op": "Eq"
[2022-11-08T07:52:47.555Z] }
[2022-11-08T07:52:47.555Z] ],
[2022-11-08T07:52:47.555Z] "config": "ewogICJqb2ItbmFtZSI6ICJ0ZXN0LXNlbGVjdG9ycyIsCiAgIndvcmtlci1jb3VudCI6IDQsCiAgInRhcmdldC10aWNrIjogMTAwLAogICJldGNkLXdhdGNoLWVuYWJsZSI6IHRydWUsCiAgImV0Y2QtZW5kcG9pbnRzIjogWyJldGNkLXN0YW5kYWxvbmU6MjM3OSJdLAogICJldGNkLXdhdGNoLXByZWZpeCI6ICIvZmFrZS1qb2IvdGVzdC8iCn0K"
[2022-11-08T07:52:47.555Z] }
[2022-11-08T07:52:47.555Z] % Total % Received % Xferd Average Speed Time Time Time Current
[2022-11-08T07:52:47.555Z] Dload Upload Total Spent Left Speed
[2022-11-08T07:52:47.555Z]
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 399 0 0 100 399 0 553k --:--:-- --:--:-- --:--:-- 389k
[2022-11-08T07:52:47.555Z] curl: (56) Recv failure: Connection reset by peer
[2022-11-08T07:52:47.555Z] job_id:
[2022-11-08T07:52:47.556Z] will execute with retry, cmd: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
[2022-11-08T07:54:33.206Z] retry 100 time(s), cmd: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
[2022-11-08T07:54:33.206Z] % Total % Received % Xferd Average Speed Time Time Time Current
[2022-11-08T07:54:33.206Z] Dload Upload Total Spent Left Speed
[2022-11-08T07:54:33.206Z]
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 62 100 {"code":"DFLOW:ErrJobNotFound","message":"job is not found"}
[2022-11-08T07:54:33.206Z] 62 0 0 20557 0 --:--:-- --:--:-- --:--:-- 31000
[2022-11-08T07:54:33.465Z] false
[2022-11-08T07:54:33.465Z]
[2022-11-08T07:54:33.465Z] failed to execute cmd for 100 times: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
```
### Anything else we need to know
- Does this test exist for other branches as well?
- Has there been a high frequency of failure lately? | 1.0 | Unstable integration test e2e_with_selectors - ### Which jobs are flaking?
dataflow engine integration test
### Which test(s) are flaking?
e2e_with_selectors
### Jenkins logs or GitHub Actions link
https://ci2.pingcap.net/blue/organizations/jenkins/engine_ghpr_integration_test/detail/engine_ghpr_integration_test/1774/pipeline
detail logs:
[log-e2e_with_selectors-dm_many_tables.tar.gz](https://github.com/pingcap/tiflow/files/9959629/log-e2e_with_selectors-dm_many_tables.tar.gz)
```go
[2022-11-08T07:52:47.555Z] create_job_json: {
[2022-11-08T07:52:47.555Z] "type": "FakeJob",
[2022-11-08T07:52:47.555Z] "selectors": [
[2022-11-08T07:52:47.555Z] {
[2022-11-08T07:52:47.555Z] "label": "name",
[2022-11-08T07:52:47.555Z] "target": "exec-1",
[2022-11-08T07:52:47.555Z] "op": "Eq"
[2022-11-08T07:52:47.555Z] }
[2022-11-08T07:52:47.555Z] ],
[2022-11-08T07:52:47.555Z] "config": "ewogICJqb2ItbmFtZSI6ICJ0ZXN0LXNlbGVjdG9ycyIsCiAgIndvcmtlci1jb3VudCI6IDQsCiAgInRhcmdldC10aWNrIjogMTAwLAogICJldGNkLXdhdGNoLWVuYWJsZSI6IHRydWUsCiAgImV0Y2QtZW5kcG9pbnRzIjogWyJldGNkLXN0YW5kYWxvbmU6MjM3OSJdLAogICJldGNkLXdhdGNoLXByZWZpeCI6ICIvZmFrZS1qb2IvdGVzdC8iCn0K"
[2022-11-08T07:52:47.555Z] }
[2022-11-08T07:52:47.555Z] % Total % Received % Xferd Average Speed Time Time Time Current
[2022-11-08T07:52:47.555Z] Dload Upload Total Spent Left Speed
[2022-11-08T07:52:47.555Z]
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 399 0 0 100 399 0 553k --:--:-- --:--:-- --:--:-- 389k
[2022-11-08T07:52:47.555Z] curl: (56) Recv failure: Connection reset by peer
[2022-11-08T07:52:47.555Z] job_id:
[2022-11-08T07:52:47.556Z] will execute with retry, cmd: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
[2022-11-08T07:54:33.206Z] retry 100 time(s), cmd: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
[2022-11-08T07:54:33.206Z] % Total % Received % Xferd Average Speed Time Time Time Current
[2022-11-08T07:54:33.206Z] Dload Upload Total Spent Left Speed
[2022-11-08T07:54:33.206Z]
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 62 100 {"code":"DFLOW:ErrJobNotFound","message":"job is not found"}
[2022-11-08T07:54:33.206Z] 62 0 0 20557 0 --:--:-- --:--:-- --:--:-- 31000
[2022-11-08T07:54:33.465Z] false
[2022-11-08T07:54:33.465Z]
[2022-11-08T07:54:33.465Z] failed to execute cmd for 100 times: curl "http://127.0.0.1:10245/api/v1/jobs/" | tee /dev/stderr | jq -e '.state == "Finished"'
```
### Anything else we need to know
- Does this test exist for other branches as well?
- Has there been a high frequency of failure lately? | test | unstable integration test with selectors which jobs are flaking dataflow engine integration test which test s are flaking with selectors jenkins logs or github actions link detail logs go create job json type fakejob selectors label name target exec op eq config total received xferd average speed time time time current dload upload total spent left speed curl recv failure connection reset by peer job id will execute with retry cmd curl tee dev stderr jq e state finished retry time s cmd curl tee dev stderr jq e state finished total received xferd average speed time time time current dload upload total spent left speed code dflow errjobnotfound message job is not found false failed to execute cmd for times curl tee dev stderr jq e state finished anything else we need to know does this test exist for other branches as well has there been a high frequency of failure lately | 1 |
33,845 | 4,863,598,160 | IssuesEvent | 2016-11-14 15:51:22 | Azure/azure-event-hubs-dotnet | https://api.github.com/repos/Azure/azure-event-hubs-dotnet | closed | Tests execution are done for a single target only | unit-tests | The client is targeting both .NET Standard 1.3 and .NET Framework 4.5.1, but tests are executed against .NET Starndard 1.3 only.

| 1.0 | Tests execution are done for a single target only - The client is targeting both .NET Standard 1.3 and .NET Framework 4.5.1, but tests are executed against .NET Starndard 1.3 only.

| test | tests execution are done for a single target only the client is targeting both net standard and net framework but tests are executed against net starndard only | 1 |
24,721 | 24,190,420,194 | IssuesEvent | 2022-09-23 16:58:18 | penrose/penrose | https://api.github.com/repos/penrose/penrose | closed | The IDE doesn't always show the whole diagram | kind:bug kind:usability system:ide small-but-annoying | Depending on how the window is sized, the `Diagram` pane of the IDE might not stretch to show the whole diagram.
For example, here's one window size where the diagram is cut off at the bottom:

Here's another window size where you can see the whole thing:

This behavior can be misleading—here, for instance, I thought there was a problem with my `onCanvas` constraints! In the top example you can still scroll down to see the rest of the diagram, but that's not immediately clear, and is fairly annoying.
**Proposed fix:** scale the diagram pane so that it always shows the whole diagram, independent of window size. | True | The IDE doesn't always show the whole diagram - Depending on how the window is sized, the `Diagram` pane of the IDE might not stretch to show the whole diagram.
For example, here's one window size where the diagram is cut off at the bottom:

Here's another window size where you can see the whole thing:

This behavior can be misleading—here, for instance, I thought there was a problem with my `onCanvas` constraints! In the top example you can still scroll down to see the rest of the diagram, but that's not immediately clear, and is fairly annoying.
**Proposed fix:** scale the diagram pane so that it always shows the whole diagram, independent of window size. | non_test | the ide doesn t always show the whole diagram depending on how the window is sized the diagram pane of the ide might not stretch to show the whole diagram for example here s one window size where the diagram is cut off at the bottom here s another window size where you can see the whole thing this behavior can be misleading—here for instance i thought there was a problem with my oncanvas constraints in the top example you can still scroll down to see the rest of the diagram but that s not immediately clear and is fairly annoying proposed fix scale the diagram pane so that it always shows the whole diagram independent of window size | 0 |
114,190 | 9,692,380,663 | IssuesEvent | 2019-05-24 13:41:32 | cselab/YMeRo | https://api.github.com/repos/cselab/YMeRo | closed | spring forces in rod | feature missing test | make independent parameters along line and crosses
make constants independent of resolution | 1.0 | spring forces in rod - make independent parameters along line and crosses
make constants independent of resolution | test | spring forces in rod make independent parameters along line and crosses make constants independent of resolution | 1 |
261,433 | 8,231,492,909 | IssuesEvent | 2018-09-07 16:03:12 | bitrockteam/databiz-website | https://api.github.com/repos/bitrockteam/databiz-website | opened | Templating & render | Priority: Low Status: Available Type: Enhancement Type: Maintenance | Following from #2 , the website will then need to be based on a dynamic template engine to take advantage of the decoupling from the content and reducing source code.
[Lit-html](https://polymer.github.io/lit-html/) will be the ideal solution since we will only need pure aesthetic and stateless components.
For performance and SEO best practices it will be **rendered sever-side** by [Puppeteer](https://github.com/GoogleChrome/puppeteer) on build time. | 1.0 | Templating & render - Following from #2 , the website will then need to be based on a dynamic template engine to take advantage of the decoupling from the content and reducing source code.
[Lit-html](https://polymer.github.io/lit-html/) will be the ideal solution since we will only need pure aesthetic and stateless components.
For performance and SEO best practices it will be **rendered sever-side** by [Puppeteer](https://github.com/GoogleChrome/puppeteer) on build time. | non_test | templating render following from the website will then need to be based on a dynamic template engine to take advantage of the decoupling from the content and reducing source code will be the ideal solution since we will only need pure aesthetic and stateless components for performance and seo best practices it will be rendered sever side by on build time | 0 |
12,470 | 2,700,690,634 | IssuesEvent | 2015-04-04 13:12:11 | CallumKeleher/multitheftauto | https://api.github.com/repos/CallumKeleher/multitheftauto | closed | Optimized PNGs | auto-migrated Priority-Medium Type-Defect | ```
As the previous patch, this one contains optimized PNGs too. Now for MTA
core distribution.
P.S. this patch is for compiled MTA:SA as I couldn't find the PNG resources
in MTA source.
```
Original issue reported on code.google.com by `skybon@live.ru` on 16 Apr 2009 at 5:30
Attachments:
* [mta-pngmod-unpackpatch.7z](https://storage.googleapis.com/google-code-attachments/multitheftauto/issue-1/comment-0/mta-pngmod-unpackpatch.7z)
| 1.0 | Optimized PNGs - ```
As the previous patch, this one contains optimized PNGs too. Now for MTA
core distribution.
P.S. this patch is for compiled MTA:SA as I couldn't find the PNG resources
in MTA source.
```
Original issue reported on code.google.com by `skybon@live.ru` on 16 Apr 2009 at 5:30
Attachments:
* [mta-pngmod-unpackpatch.7z](https://storage.googleapis.com/google-code-attachments/multitheftauto/issue-1/comment-0/mta-pngmod-unpackpatch.7z)
| non_test | optimized pngs as the previous patch this one contains optimized pngs too now for mta core distribution p s this patch is for compiled mta sa as i couldn t find the png resources in mta source original issue reported on code google com by skybon live ru on apr at attachments | 0 |
24,946 | 4,118,457,680 | IssuesEvent | 2016-06-08 11:34:03 | mathjax/MathJax | https://api.github.com/repos/mathjax/MathJax | closed | Line breaking inside shrink-wrap elements doesn't work well | Accepted Merged Test Needed | For elements that shrink-wrap their contents (like table cells or inline-block elements), the container width is measured as 0, and so line breaking can't be properly done.
Find a way to get the proper container width in that case (while still handling floating elements and other things that affect the width).
See issue #571, and also #1466. | 1.0 | Line breaking inside shrink-wrap elements doesn't work well - For elements that shrink-wrap their contents (like table cells or inline-block elements), the container width is measured as 0, and so line breaking can't be properly done.
Find a way to get the proper container width in that case (while still handling floating elements and other things that affect the width).
See issue #571, and also #1466. | test | line breaking inside shrink wrap elements doesn t work well for elements that shrink wrap their contents like table cells or inline block elements the container width is measured as and so line breaking can t be properly done find a way to get the proper container width in that case while still handling floating elements and other things that affect the width see issue and also | 1 |
423,523 | 12,298,084,786 | IssuesEvent | 2020-05-11 09:54:57 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | interdato.000webhostapp.com - site is not usable | browser-fenix engine-gecko priority-normal | <!-- @browser: Firefox Mobile 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:78.0) Gecko/78.0 Firefox/78.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52647 -->
<!-- @extra_labels: browser-fenix -->
**URL**: http://interdato.000webhostapp.com/
**Browser / Version**: Firefox Mobile 78.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
You can only see 2 buttons and a loading message
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | interdato.000webhostapp.com - site is not usable - <!-- @browser: Firefox Mobile 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:78.0) Gecko/78.0 Firefox/78.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52647 -->
<!-- @extra_labels: browser-fenix -->
**URL**: http://interdato.000webhostapp.com/
**Browser / Version**: Firefox Mobile 78.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
You can only see 2 buttons and a loading message
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | interdato com site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description page not loading correctly steps to reproduce you can only see buttons and a loading message browser configuration none from with ❤️ | 0 |
151,876 | 12,061,444,790 | IssuesEvent | 2020-04-15 23:49:25 | Niraj-Kamdar/question-paper-generator | https://api.github.com/repos/Niraj-Kamdar/question-paper-generator | closed | unittest for users class | Test | - [x] unittest for login
- [x] unittest for account
- [x] unittest for reset password
- [x] unittest for reset token | 1.0 | unittest for users class - - [x] unittest for login
- [x] unittest for account
- [x] unittest for reset password
- [x] unittest for reset token | test | unittest for users class unittest for login unittest for account unittest for reset password unittest for reset token | 1 |
103,120 | 8,880,360,989 | IssuesEvent | 2019-01-14 05:32:54 | flemingcaleb/InfraBot | https://api.github.com/repos/flemingcaleb/InfraBot | closed | Add command for help | Pushed To Testing enhancement | Add a !help command that lists all available commands and a short blurb about them. | 1.0 | Add command for help - Add a !help command that lists all available commands and a short blurb about them. | test | add command for help add a help command that lists all available commands and a short blurb about them | 1 |
92,360 | 26,664,322,848 | IssuesEvent | 2023-01-26 01:03:33 | runatlantis/atlantis | https://api.github.com/repos/runatlantis/atlantis | closed | Automate releases without a manual PR prior to release | build | <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you!
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request.
- If you are interested in working on this issue or have submitted a pull request, please leave a comment.
<!--- Thank you for keeping this note for the community --->
---
- [x] I'd be willing to implement this feature ([contributing guide](https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md))
**Describe the user story**
<!--
A clear and concise description of what workflow is meant to be improved.
Example: "As a developer, I often want to do <something>, but I often face <problem>".
-->
I'd like to automate releases without making manual changes
- to avoid issues like this
- https://github.com/runatlantis/atlantis/issues/2909
- to avoid chore prs like
- https://github.com/runatlantis/atlantis/pull/3005
- #2910
- #2734
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen. Consider that atlantis is used
by many people, and your particular use case might not make sense to implement in the core.
-->
Instead of hard coding three version, let's pass in the version using an ldflag via goreleaser
See https://github.com/cloudposse/atmos/blob/3a43ea01e4704705a23afae861931e07bafdfc8d/cmd/version.go#L8 (their default is `0.0.1`, ours can be `build` or `local`) and https://github.com/cloudposse/atmos/blob/3a43ea01e4704705a23afae861931e07bafdfc8d/.goreleaser.yml#L27 for inspiration.
For the kustomize bundle, let's just use `latest` since that is already tied to the latest stable release.
Scenarios after completion
- PR merge - should result in version `dev-<7 char git ref>`.
- prerelease and release - results in version with same tag name
Currently above scenarios will return hard coded value from code
Here is where we currently document our current manual release pr steps
https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md#creating-a-new-release
**Describe the drawbacks of your solution**
<!--
This section is important not only to identify future issues, but also for us to see whether
you thought through your request. When filling it, ask yourself what are the problems we could
have maintaining what you propose. How often will it break?
-->
None
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered,
and why you think they wouldn't be good enough.
-->
None
- This may be a duplicate or related to https://github.com/runatlantis/atlantis/issues/1461 | 1.0 | Automate releases without a manual PR prior to release - <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you!
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request.
- If you are interested in working on this issue or have submitted a pull request, please leave a comment.
<!--- Thank you for keeping this note for the community --->
---
- [x] I'd be willing to implement this feature ([contributing guide](https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md))
**Describe the user story**
<!--
A clear and concise description of what workflow is meant to be improved.
Example: "As a developer, I often want to do <something>, but I often face <problem>".
-->
I'd like to automate releases without making manual changes
- to avoid issues like this
- https://github.com/runatlantis/atlantis/issues/2909
- to avoid chore prs like
- https://github.com/runatlantis/atlantis/pull/3005
- #2910
- #2734
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen. Consider that atlantis is used
by many people, and your particular use case might not make sense to implement in the core.
-->
Instead of hard coding three version, let's pass in the version using an ldflag via goreleaser
See https://github.com/cloudposse/atmos/blob/3a43ea01e4704705a23afae861931e07bafdfc8d/cmd/version.go#L8 (their default is `0.0.1`, ours can be `build` or `local`) and https://github.com/cloudposse/atmos/blob/3a43ea01e4704705a23afae861931e07bafdfc8d/.goreleaser.yml#L27 for inspiration.
For the kustomize bundle, let's just use `latest` since that is already tied to the latest stable release.
Scenarios after completion
- PR merge - should result in version `dev-<7 char git ref>`.
- prerelease and release - results in version with same tag name
Currently above scenarios will return hard coded value from code
Here is where we currently document our current manual release pr steps
https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md#creating-a-new-release
**Describe the drawbacks of your solution**
<!--
This section is important not only to identify future issues, but also for us to see whether
you thought through your request. When filling it, ask yourself what are the problems we could
have maintaining what you propose. How often will it break?
-->
None
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered,
and why you think they wouldn't be good enough.
-->
None
- This may be a duplicate or related to https://github.com/runatlantis/atlantis/issues/1461 | non_test | automate releases without a manual pr prior to release community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request searching for pre existing feature requests helps us consolidate datapoints for identical requirements into a single place thank you please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment i d be willing to implement this feature describe the user story a clear and concise description of what workflow is meant to be improved example as a developer i often want to do but i often face i d like to automate releases without making manual changes to avoid issues like this to avoid chore prs like describe the solution you d like a clear and concise description of what you want to happen consider that atlantis is used by many people and your particular use case might not make sense to implement in the core instead of hard coding three version let s pass in the version using an ldflag via goreleaser see their default is ours can be build or local and for inspiration for the kustomize bundle let s just use latest since that is already tied to the latest stable release scenarios after completion pr merge should result in version dev prerelease and release results in version with same tag name currently above scenarios will return hard coded value from code here is where we currently document our current manual release pr steps describe the drawbacks of your solution this section is important not only to identify future issues but also for us to see whether you thought through your request when filling it ask yourself what are the problems we could have maintaining what you propose how often will it break none describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered and why you think they wouldn t be good enough none this may be a duplicate or related to | 0 |
40,541 | 12,799,561,178 | IssuesEvent | 2020-07-02 15:35:09 | TreyM-WSS/WhiteSource-Demo | https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo | opened | CVE-2018-12022 (High) detected in jackson-databind-2.8.1.jar | security vulnerability | ## CVE-2018-12022 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4, 2.8.11.2, 2.9.6"}],"vulnerabilityIdentifier":"CVE-2018-12022","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-12022 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2018-12022 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4, 2.8.11.2, 2.9.6"}],"vulnerabilityIdentifier":"CVE-2018-12022","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm whitesource demo pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details an issue was discovered in fasterxml jackson databind prior to and when default typing is enabled either globally or for a specific property the service has the jodd db jar for database access for the jodd framework in the classpath and an attacker can provide an ldap service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in fasterxml jackson databind prior to and when default typing is enabled either globally or for a specific property the service has the jodd db jar for database access for the jodd framework in the classpath and an attacker can provide an ldap service to access it is possible to make the service execute a malicious payload vulnerabilityurl | 0 |
31,569 | 4,710,803,098 | IssuesEvent | 2016-10-14 11:34:52 | outscale/butterfly | https://api.github.com/repos/outscale/butterfly | closed | tests: check that sg rules cumulates | tests | Initial setup:
- VM1 on sg1
- VM2 (no sg)
- sg1 has one rule allowing UDP port 8000
- sg2 has one rule allowing UDP port 9000
Test that:
- UDP communication VM2 --> VM1 on port 8000 is OK
- UDP communication VM2 --> VM1 on port 9000 is KO
Change setup:
- VM1 on sg1 and sg2
Test that:
- UDP communication VM2 --> VM1 on port 8000 is OK
- UDP communication VM2 --> VM1 on port 9000 is OK
Change setup:
- VM1 on sg2
Test that:
- UDP communication VM2 --> VM1 on port 8000 is KO
- UDP communication VM2 --> VM1 on port 9000 is OK
Change setup:
- VM1 has no sg
Test that:
- UDP communication VM2 --> VM1 on port 8000 is KO
- UDP communication VM2 --> VM1 on port 9000 is KO | 1.0 | tests: check that sg rules cumulates - Initial setup:
- VM1 on sg1
- VM2 (no sg)
- sg1 has one rule allowing UDP port 8000
- sg2 has one rule allowing UDP port 9000
Test that:
- UDP communication VM2 --> VM1 on port 8000 is OK
- UDP communication VM2 --> VM1 on port 9000 is KO
Change setup:
- VM1 on sg1 and sg2
Test that:
- UDP communication VM2 --> VM1 on port 8000 is OK
- UDP communication VM2 --> VM1 on port 9000 is OK
Change setup:
- VM1 on sg2
Test that:
- UDP communication VM2 --> VM1 on port 8000 is KO
- UDP communication VM2 --> VM1 on port 9000 is OK
Change setup:
- VM1 has no sg
Test that:
- UDP communication VM2 --> VM1 on port 8000 is KO
- UDP communication VM2 --> VM1 on port 9000 is KO | test | tests check that sg rules cumulates initial setup on no sg has one rule allowing udp port has one rule allowing udp port test that udp communication on port is ok udp communication on port is ko change setup on and test that udp communication on port is ok udp communication on port is ok change setup on test that udp communication on port is ko udp communication on port is ok change setup has no sg test that udp communication on port is ko udp communication on port is ko | 1 |
98,423 | 8,676,626,415 | IssuesEvent | 2018-11-30 14:39:32 | OHDSI/Atlas | https://api.github.com/repos/OHDSI/Atlas | closed | loading indicator is not visible | bug data sources fixed - needs review/testing | ### Expected behavior
user always is aware when loading is taking place
### Actual behavior
on some screen resolutions or if an existing report is very tall (height) the loading indicator appears below the fold (off the screen) and is not visible
| 1.0 | loading indicator is not visible - ### Expected behavior
user always is aware when loading is taking place
### Actual behavior
on some screen resolutions or if an existing report is very tall (height) the loading indicator appears below the fold (off the screen) and is not visible
| test | loading indicator is not visible expected behavior user always is aware when loading is taking place actual behavior on some screen resolutions or if an existing report is very tall height the loading indicator appears below the fold off the screen and is not visible | 1 |
9,924 | 12,963,469,714 | IssuesEvent | 2020-07-20 18:50:31 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | reopened | misleading categorization - azure devops server 2019 does not support parameters | Pri2 devops-cicd-process/tech devops/prod doc-enhancement | The selector in the upper left corner is misleading users - it is possible to select server 2019 but according to https://docs.microsoft.com/en-us/azure/devops/release-notes/features-timeline it will only be supported in server 2020
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 790318bb-8220-3241-4ca7-73351074492f
* Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86
* Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script)
* Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | misleading categorization - azure devops server 2019 does not support parameters - The selector in the upper left corner is misleading users - it is possible to select server 2019 but according to https://docs.microsoft.com/en-us/azure/devops/release-notes/features-timeline it will only be supported in server 2020
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 790318bb-8220-3241-4ca7-73351074492f
* Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86
* Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script)
* Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | non_test | misleading categorization azure devops server does not support parameters the selector in the upper left corner is misleading users it is possible to select server but according to it will only be supported in server document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 0 |
22,576 | 19,678,432,687 | IssuesEvent | 2022-01-11 14:38:41 | hl7ch/ch-allergyintolerance | https://api.github.com/repos/hl7ch/ch-allergyintolerance | closed | Besseren Zugang zu den einzelnen Seiten des IG ermöglichen (Emmanuel Eschmann, CISTEC AG) | STU 1 Ballot usability | ch.fhir.ig.ch-allergyintolerance#0.2.0 /
Die Headerleiste (weisse Schrift auf roten Grund) führt nur die beiden Seiten "Home" und "Artifacts" auf. Profile, Extensions, Terminologien und Beispielen werden gemeinsam auf der Seite "Artifacts" aufgelistet.
Vorschlag: Wie bei den anderen CH-IG's die Seiten "Home", "Profiles", "Extensions", "Terminology" und "Artifacts" in der Headerleiste aufführen und Profile, Extensions und die Terminologien in jeweils eigenen Seiten detaillieren.
Emmanuel Eschmann, CISTEC AG
| True | Besseren Zugang zu den einzelnen Seiten des IG ermöglichen (Emmanuel Eschmann, CISTEC AG) - ch.fhir.ig.ch-allergyintolerance#0.2.0 /
Die Headerleiste (weisse Schrift auf roten Grund) führt nur die beiden Seiten "Home" und "Artifacts" auf. Profile, Extensions, Terminologien und Beispielen werden gemeinsam auf der Seite "Artifacts" aufgelistet.
Vorschlag: Wie bei den anderen CH-IG's die Seiten "Home", "Profiles", "Extensions", "Terminology" und "Artifacts" in der Headerleiste aufführen und Profile, Extensions und die Terminologien in jeweils eigenen Seiten detaillieren.
Emmanuel Eschmann, CISTEC AG
| non_test | besseren zugang zu den einzelnen seiten des ig ermöglichen emmanuel eschmann cistec ag ch fhir ig ch allergyintolerance die headerleiste weisse schrift auf roten grund führt nur die beiden seiten home und artifacts auf profile extensions terminologien und beispielen werden gemeinsam auf der seite artifacts aufgelistet vorschlag wie bei den anderen ch ig s die seiten home profiles extensions terminology und artifacts in der headerleiste aufführen und profile extensions und die terminologien in jeweils eigenen seiten detaillieren emmanuel eschmann cistec ag | 0 |
536,783 | 15,713,223,751 | IssuesEvent | 2021-03-27 15:19:16 | GoogleContainerTools/kaniko | https://api.github.com/repos/GoogleContainerTools/kaniko | closed | Pushing same image version into Artifactory gets INVALID_MANIFEST error | area/behavior kind/bug priority/p3 | Hi,
when using the executor and pushing same image version again into JFrog Artifactory I get error.
this doesn't happen when using docker though.
**Actual behavior**
error pushing image: failed to push to destination artifactory:5001/action_server:1.1.0: MANIFEST_INVALID: manifest invalid; map[description:HTTP 403 Forbidden]
**The error looks like Authorization, but it is not. changing the image version to a new version will go successful.**
**Expected behavior**
the push to the artifactory to be successful. Artifactory allows overriding the same image version, since we're doing the same with docker.
**Triage Notes for the Maintainers**
<!-- 🎉🎉🎉 Thank you for an opening an issue !!! 🎉🎉🎉
We are doing our best to get to this. Please help us by helping us prioritize your issue by filling the section below -->
| **Description** | **Yes/No** |
|----------------|---------------|
| Please check if this a new feature you are proposing | <ul><li>- [ ] </li></ul>|
| Please check if the build works in docker but not in kaniko | <ul><li>- [X] </li></ul>|
| Please check if this error is seen when you use `--cache` flag | <ul><li>- [ ] </li></ul>|
| Please check if your dockerfile is a multistage dockerfile | <ul><li>- [ ] </li></ul>|
| 1.0 | Pushing same image version into Artifactory gets INVALID_MANIFEST error - Hi,
when using the executor and pushing same image version again into JFrog Artifactory I get error.
this doesn't happen when using docker though.
**Actual behavior**
error pushing image: failed to push to destination artifactory:5001/action_server:1.1.0: MANIFEST_INVALID: manifest invalid; map[description:HTTP 403 Forbidden]
**The error looks like Authorization, but it is not. changing the image version to a new version will go successful.**
**Expected behavior**
the push to the artifactory to be successful. Artifactory allows overriding the same image version, since we're doing the same with docker.
**Triage Notes for the Maintainers**
<!-- 🎉🎉🎉 Thank you for an opening an issue !!! 🎉🎉🎉
We are doing our best to get to this. Please help us by helping us prioritize your issue by filling the section below -->
| **Description** | **Yes/No** |
|----------------|---------------|
| Please check if this a new feature you are proposing | <ul><li>- [ ] </li></ul>|
| Please check if the build works in docker but not in kaniko | <ul><li>- [X] </li></ul>|
| Please check if this error is seen when you use `--cache` flag | <ul><li>- [ ] </li></ul>|
| Please check if your dockerfile is a multistage dockerfile | <ul><li>- [ ] </li></ul>|
| non_test | pushing same image version into artifactory gets invalid manifest error hi when using the executor and pushing same image version again into jfrog artifactory i get error this doesn t happen when using docker though actual behavior error pushing image failed to push to destination artifactory action server manifest invalid manifest invalid map the error looks like authorization but it is not changing the image version to a new version will go successful expected behavior the push to the artifactory to be successful artifactory allows overriding the same image version since we re doing the same with docker triage notes for the maintainers 🎉🎉🎉 thank you for an opening an issue 🎉🎉🎉 we are doing our best to get to this please help us by helping us prioritize your issue by filling the section below description yes no please check if this a new feature you are proposing please check if the build works in docker but not in kaniko please check if this error is seen when you use cache flag please check if your dockerfile is a multistage dockerfile | 0 |
77,461 | 3,506,389,426 | IssuesEvent | 2016-01-08 06:23:21 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | opened | Crash OregonCore.exe (BB #506) | Category: Crash migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** maikale
**Original Date:** 26.02.2014 14:47:07 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** on hold
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/506
<hr>
Hello today I compiled over 10 times OregonCore Sorce and every time I Crash Oregon.exe
if can anyone help me would be very grateful thanks | 1.0 | Crash OregonCore.exe (BB #506) - This issue was migrated from bitbucket.
**Original Reporter:** maikale
**Original Date:** 26.02.2014 14:47:07 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** on hold
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/506
<hr>
Hello today I compiled over 10 times OregonCore Sorce and every time I Crash Oregon.exe
if can anyone help me would be very grateful thanks | non_test | crash oregoncore exe bb this issue was migrated from bitbucket original reporter maikale original date gmt original priority major original type bug original state on hold direct link hello today i compiled over times oregoncore sorce and every time i crash oregon exe if can anyone help me would be very grateful thanks | 0 |
112,553 | 9,583,512,606 | IssuesEvent | 2019-05-08 05:48:32 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Running unit tests in release mode fails when not official build | QA/No QA/Test-Plan-Specified release-notes/exclude | ## Description
When building from source, specifically a release build with official_build=false, there are two test failures:
```
2 tests failed:
MediaYouTubeTest.GetChannelUrl (../../brave/vendor/bat-native-ledger/src/bat/ledger/internal/media/youtube_unittest.cc:498)
MediaYouTubeTest.GetVideoUrl (../../brave/vendor/bat-native-ledger/src/bat/ledger/internal/media/youtube_unittest.cc:466)
```
Problem introduced in https://github.com/brave/brave-core/pull/2167, meaning it occurs in master/0.66.x/0.65.x
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clone `brave-browser`
2. `npm run init -- --all && npm run build -- Release --official_build=false`
3. `npm test brave_unit_tests -- Release`
## Actual result:
Errors
## Expected result:
Tests should pass
| 1.0 | Running unit tests in release mode fails when not official build - ## Description
When building from source, specifically a release build with official_build=false, there are two test failures:
```
2 tests failed:
MediaYouTubeTest.GetChannelUrl (../../brave/vendor/bat-native-ledger/src/bat/ledger/internal/media/youtube_unittest.cc:498)
MediaYouTubeTest.GetVideoUrl (../../brave/vendor/bat-native-ledger/src/bat/ledger/internal/media/youtube_unittest.cc:466)
```
Problem introduced in https://github.com/brave/brave-core/pull/2167, meaning it occurs in master/0.66.x/0.65.x
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clone `brave-browser`
2. `npm run init -- --all && npm run build -- Release --official_build=false`
3. `npm test brave_unit_tests -- Release`
## Actual result:
Errors
## Expected result:
Tests should pass
| test | running unit tests in release mode fails when not official build description when building from source specifically a release build with official build false there are two test failures tests failed mediayoutubetest getchannelurl brave vendor bat native ledger src bat ledger internal media youtube unittest cc mediayoutubetest getvideourl brave vendor bat native ledger src bat ledger internal media youtube unittest cc problem introduced in meaning it occurs in master x x steps to reproduce clone brave browser npm run init all npm run build release official build false npm test brave unit tests release actual result errors expected result tests should pass | 1 |
331,074 | 24,290,299,788 | IssuesEvent | 2022-09-29 05:00:12 | dankrzeminski32/BirthdayDiscordBot | https://api.github.com/repos/dankrzeminski32/BirthdayDiscordBot | closed | Organize + Project Clean up | documentation | Current situation:
REAME.md is messy.
Re-used code in most functions, make reference methods.
Expected situation/ideas:
Reduce file size's
Have the project organized(get rid of useless classes/files)
No more un used imports. | 1.0 | Organize + Project Clean up - Current situation:
REAME.md is messy.
Re-used code in most functions, make reference methods.
Expected situation/ideas:
Reduce file size's
Have the project organized(get rid of useless classes/files)
No more un used imports. | non_test | organize project clean up current situation reame md is messy re used code in most functions make reference methods expected situation ideas reduce file size s have the project organized get rid of useless classes files no more un used imports | 0 |
273,700 | 23,778,918,704 | IssuesEvent | 2022-09-02 01:02:57 | AtlasOfLivingAustralia/profile-hub | https://api.github.com/repos/AtlasOfLivingAustralia/profile-hub | closed | Map filter bug if exclude is used? | Testing complete ABRS priority 1 | There is an problem with the 'Edit Map Configuration' feature when editing a profile. It seems to occur when facets are excluded using the 'Choose More' option.
This issue can be recreated by:
- Selecting the Acacia buxifolia profile in the MLH test collection on profiles-dev,
- Limiting the taxonomy to only the Acacia buxifolia species.
- Excluding all records with country = unknown (Note: this displays two lines under the Selected filters which is an error).
- Trying to include only “Goobang” option for Capad 2014 terrestrial facet...
At which point all the points disappear and the Selected filters haven’t updated... I also can’t select additional options... the only option that seems to work is to clear the filters and start again... interestingly, after a short while we get a couple of error messages which may be related.
| 1.0 | Map filter bug if exclude is used? - There is an problem with the 'Edit Map Configuration' feature when editing a profile. It seems to occur when facets are excluded using the 'Choose More' option.
This issue can be recreated by:
- Selecting the Acacia buxifolia profile in the MLH test collection on profiles-dev,
- Limiting the taxonomy to only the Acacia buxifolia species.
- Excluding all records with country = unknown (Note: this displays two lines under the Selected filters which is an error).
- Trying to include only “Goobang” option for Capad 2014 terrestrial facet...
At which point all the points disappear and the Selected filters haven’t updated... I also can’t select additional options... the only option that seems to work is to clear the filters and start again... interestingly, after a short while we get a couple of error messages which may be related.
| test | map filter bug if exclude is used there is an problem with the edit map configuration feature when editing a profile it seems to occur when facets are excluded using the choose more option this issue can be recreated by selecting the acacia buxifolia profile in the mlh test collection on profiles dev limiting the taxonomy to only the acacia buxifolia species excluding all records with country unknown note this displays two lines under the selected filters which is an error trying to include only “goobang” option for capad terrestrial facet at which point all the points disappear and the selected filters haven’t updated i also can’t select additional options the only option that seems to work is to clear the filters and start again interestingly after a short while we get a couple of error messages which may be related | 1 |
22,536 | 7,187,947,329 | IssuesEvent | 2018-02-02 08:13:23 | FreeRDP/FreeRDP | https://api.github.com/repos/FreeRDP/FreeRDP | closed | Nigthly build fails, seems related to PR 4328 | build | Today I tried to build latest nightly build on rpi 3 with debian Stretch but it fails:
`
...
[100%] Generating xfreerdp.1
cd "/home/pi/Downloads/freerdp-nightly-2.0.0+0~20180124024836.476~1.gbp32cc6e/obj-arm-linux-gnueabihf/client/X11" && "/home/pi/Downloads/freerdp-nightly-2.0.0+0~20180124024836.476~1.gbp32cc6e/obj-arm-linux-gnueabihf/client/X11/generate_argument_docbook"
==6324==ASan runtime does not come first in initial library list; you should either link runtime to your application or manually preload it with LD_PRELOAD.
client/X11/CMakeFiles/xfreerdp.manpage.dir/build.make:67: set di istruzioni per l'obiettivo "client/X11/xfreerdp.1" non riuscito
make[3]: *** [client/X11/xfreerdp.1] Errore 1
`
I suppose is related to the commit 32cc6e16ef6e7148dac90502cb074366075f820d that probably need further additions.
From a fast search I not found other solution that manually add libasan path to LD_PRELOAD on debian/rules that FWIK is not a good idea.
Thanks for any reply and sorry for my bad english. | 1.0 | Nigthly build fails, seems related to PR 4328 - Today I tried to build latest nightly build on rpi 3 with debian Stretch but it fails:
`
...
[100%] Generating xfreerdp.1
cd "/home/pi/Downloads/freerdp-nightly-2.0.0+0~20180124024836.476~1.gbp32cc6e/obj-arm-linux-gnueabihf/client/X11" && "/home/pi/Downloads/freerdp-nightly-2.0.0+0~20180124024836.476~1.gbp32cc6e/obj-arm-linux-gnueabihf/client/X11/generate_argument_docbook"
==6324==ASan runtime does not come first in initial library list; you should either link runtime to your application or manually preload it with LD_PRELOAD.
client/X11/CMakeFiles/xfreerdp.manpage.dir/build.make:67: set di istruzioni per l'obiettivo "client/X11/xfreerdp.1" non riuscito
make[3]: *** [client/X11/xfreerdp.1] Errore 1
`
I suppose is related to the commit 32cc6e16ef6e7148dac90502cb074366075f820d that probably need further additions.
From a fast search I not found other solution that manually add libasan path to LD_PRELOAD on debian/rules that FWIK is not a good idea.
Thanks for any reply and sorry for my bad english. | non_test | nigthly build fails seems related to pr today i tried to build latest nightly build on rpi with debian stretch but it fails generating xfreerdp cd home pi downloads freerdp nightly obj arm linux gnueabihf client home pi downloads freerdp nightly obj arm linux gnueabihf client generate argument docbook asan runtime does not come first in initial library list you should either link runtime to your application or manually preload it with ld preload client cmakefiles xfreerdp manpage dir build make set di istruzioni per l obiettivo client xfreerdp non riuscito make errore i suppose is related to the commit that probably need further additions from a fast search i not found other solution that manually add libasan path to ld preload on debian rules that fwik is not a good idea thanks for any reply and sorry for my bad english | 0 |
638,029 | 20,693,846,720 | IssuesEvent | 2022-03-11 05:18:40 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Deprecation plan for Flex Volumes? | sig/storage priority/important-longterm lifecycle/rotten triage/accepted | FAQ says `Out-of-tree FlexVolume driver [deprecated]` https://github.com/kubernetes/community/blob/master/sig-storage/volume-plugin-faq.md#in-tree-vs-out-of-tree-volume-plugins
Also this was hinted at in https://github.com/kubernetes/kubernetes/issues/78737 as well.
How and when do we roll this out?
| 1.0 | Deprecation plan for Flex Volumes? - FAQ says `Out-of-tree FlexVolume driver [deprecated]` https://github.com/kubernetes/community/blob/master/sig-storage/volume-plugin-faq.md#in-tree-vs-out-of-tree-volume-plugins
Also this was hinted at in https://github.com/kubernetes/kubernetes/issues/78737 as well.
How and when do we roll this out?
| non_test | deprecation plan for flex volumes faq says out of tree flexvolume driver also this was hinted at in as well how and when do we roll this out | 0 |
73,227 | 19,600,305,895 | IssuesEvent | 2022-01-06 00:00:43 | xamarin/xamarin-android | https://api.github.com/repos/xamarin/xamarin-android | closed | Investigate and fix MSBuild smoke test issues on Linux | Area: App+Library Build | Context: https://github.com/xamarin/xamarin-android/pull/5681#issuecomment-790017891
We should extend the test coverage of our Linux build and smoke test job. | 1.0 | Investigate and fix MSBuild smoke test issues on Linux - Context: https://github.com/xamarin/xamarin-android/pull/5681#issuecomment-790017891
We should extend the test coverage of our Linux build and smoke test job. | non_test | investigate and fix msbuild smoke test issues on linux context we should extend the test coverage of our linux build and smoke test job | 0 |
21,382 | 4,707,297,595 | IssuesEvent | 2016-10-13 19:43:07 | choderalab/yank | https://api.github.com/repos/choderalab/yank | closed | Improve usability of resuming and extending simulations | documentation enhancement | It would be great if users could easily resume or extend simulations. | 1.0 | Improve usability of resuming and extending simulations - It would be great if users could easily resume or extend simulations. | non_test | improve usability of resuming and extending simulations it would be great if users could easily resume or extend simulations | 0 |
261,507 | 22,747,779,489 | IssuesEvent | 2022-07-07 10:41:23 | zephyrproject-rtos/test_results | https://api.github.com/repos/zephyrproject-rtos/test_results | opened |
IPv4 Options Append a RTRALT option with a length of 1. error
| area: Tests |
**Describe the bug**
Append a RTRALT option with a length of 1. test is Fail on Zephyr3.0.0 on qemu_x86
**References**
RFC 791: section 3.1
RFC 792
RFC 1122: section 3.2.2 {Included octets same as received}
RFC 1122: section 3.2.2.5 {Send Parameter Problem messages}
**Results**
FAIL: icmp.v4 got [0: icmp IPv4 1300] Echo response, which is not an expected result.
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: Zephyr3.0.0
| 1.0 |
IPv4 Options Append a RTRALT option with a length of 1. error
-
**Describe the bug**
Append a RTRALT option with a length of 1. test is Fail on Zephyr3.0.0 on qemu_x86
**References**
RFC 791: section 3.1
RFC 792
RFC 1122: section 3.2.2 {Included octets same as received}
RFC 1122: section 3.2.2.5 {Send Parameter Problem messages}
**Results**
FAIL: icmp.v4 got [0: icmp IPv4 1300] Echo response, which is not an expected result.
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: Zephyr3.0.0
| test | options append a rtralt option with a length of error describe the bug append a rtralt option with a length of test is fail on on qemu references rfc section rfc rfc section included octets same as received rfc section send parameter problem messages results fail icmp got echo response which is not an expected result environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used | 1 |
581,960 | 17,348,766,803 | IssuesEvent | 2021-07-29 05:28:56 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | Subtype check for service objects with remote methods produce incorrect results | Area/RuntimeTypeChecker Priority/Blocker Team/jBallerina Type/Bug | **Description:**
$title.
Remote methods are part of the type so they need to be considered for type checking. Resource methods are not, so they can/should be ignored.
**Steps to reproduce:**
```ballerina
import ballerina/io;
service class Foo {
remote function x() {
}
}
service class Bar {
}
public function main() {
// Foo f = new Bar(); // Compilation error.
any b = new Bar();
io:println(b is Foo); // true, should be false!
}
```
**Affected Versions:**
slbeta2
**Related Issues (optional):**
https://github.com/ballerina-platform/ballerina-lang/issues/31906 | 1.0 | Subtype check for service objects with remote methods produce incorrect results - **Description:**
$title.
Remote methods are part of the type so they need to be considered for type checking. Resource methods are not, so they can/should be ignored.
**Steps to reproduce:**
```ballerina
import ballerina/io;
service class Foo {
remote function x() {
}
}
service class Bar {
}
public function main() {
// Foo f = new Bar(); // Compilation error.
any b = new Bar();
io:println(b is Foo); // true, should be false!
}
```
**Affected Versions:**
slbeta2
**Related Issues (optional):**
https://github.com/ballerina-platform/ballerina-lang/issues/31906 | non_test | subtype check for service objects with remote methods produce incorrect results description title remote methods are part of the type so they need to be considered for type checking resource methods are not so they can should be ignored steps to reproduce ballerina import ballerina io service class foo remote function x service class bar public function main foo f new bar compilation error any b new bar io println b is foo true should be false affected versions related issues optional | 0 |
99,565 | 8,704,485,314 | IssuesEvent | 2018-12-05 19:32:41 | chartjs/Chart.js | https://api.github.com/repos/chartjs/Chart.js | closed | [BUG] y-axis labels are being cut off for apparently unknown reason | status: needs test case type: bug | Looking for similar issues I found that issue [#729](https://github.com/chartjs/Chart.js/issues/729) show big numbers being cut off in the y-axis. But in my case the problem occurs when I have decimal values lower than 1.
Please take a look at my screenshot.
The first graph shows y-axis labels being cut off while the others below show them as expected.
I drawed a dashed yellow line just to show the `<canvas>` limits.

I tried set up `padding` and `width` for the `<canvas>` tag but with no success. | 1.0 | [BUG] y-axis labels are being cut off for apparently unknown reason - Looking for similar issues I found that issue [#729](https://github.com/chartjs/Chart.js/issues/729) show big numbers being cut off in the y-axis. But in my case the problem occurs when I have decimal values lower than 1.
Please take a look at my screenshot.
The first graph shows y-axis labels being cut off while the others below show them as expected.
I drawed a dashed yellow line just to show the `<canvas>` limits.

I tried set up `padding` and `width` for the `<canvas>` tag but with no success. | test | y axis labels are being cut off for apparently unknown reason looking for similar issues i found that issue show big numbers being cut off in the y axis but in my case the problem occurs when i have decimal values lower than please take a look at my screenshot the first graph shows y axis labels being cut off while the others below show them as expected i drawed a dashed yellow line just to show the limits i tried set up padding and width for the tag but with no success | 1 |
434,745 | 30,464,993,047 | IssuesEvent | 2023-07-17 09:43:53 | kodadot/developers | https://api.github.com/repos/kodadot/developers | opened | Add popular NFT related ERC standards | documentation good first issue help wanted onboarding | Would be nice to have popular NFT standards on howtonft.guide
- 721
- 998
- 1155
- 4337
- 6551
- 6381
Dynamic NFTs, onchain NFTs, what's else? | 1.0 | Add popular NFT related ERC standards - Would be nice to have popular NFT standards on howtonft.guide
- 721
- 998
- 1155
- 4337
- 6551
- 6381
Dynamic NFTs, onchain NFTs, what's else? | non_test | add popular nft related erc standards would be nice to have popular nft standards on howtonft guide dynamic nfts onchain nfts what s else | 0 |
119,928 | 10,081,910,889 | IssuesEvent | 2019-07-25 09:53:27 | zaproxy/zaproxy | https://api.github.com/repos/zaproxy/zaproxy | opened | Unit tests for BigRedirectsScanner | add-on unittest | Create unit tests for [BigRedirectsScanner](https://github.com/zaproxy/zap-extensions/blob/master/addOns/pscanrulesAlpha/src/main/java/org/zaproxy/zap/extension/pscanrulesAlpha/BigRedirectsScanner.java).
Unit tests should test all of the associated functionality and include tests for false positives.
The tests should at least check the number of alerts and the parameters and evidence (as appropriate). Additional checks can be made if thought useful.
Tests for other rules can be found in https://github.com/zaproxy/zap-extensions/tree/master/addOns/pscanrules/src/test/java/org/zaproxy/zap/extension/pscanrules | 1.0 | Unit tests for BigRedirectsScanner - Create unit tests for [BigRedirectsScanner](https://github.com/zaproxy/zap-extensions/blob/master/addOns/pscanrulesAlpha/src/main/java/org/zaproxy/zap/extension/pscanrulesAlpha/BigRedirectsScanner.java).
Unit tests should test all of the associated functionality and include tests for false positives.
The tests should at least check the number of alerts and the parameters and evidence (as appropriate). Additional checks can be made if thought useful.
Tests for other rules can be found in https://github.com/zaproxy/zap-extensions/tree/master/addOns/pscanrules/src/test/java/org/zaproxy/zap/extension/pscanrules | test | unit tests for bigredirectsscanner create unit tests for unit tests should test all of the associated functionality and include tests for false positives the tests should at least check the number of alerts and the parameters and evidence as appropriate additional checks can be made if thought useful tests for other rules can be found in | 1 |
44,535 | 5,843,584,760 | IssuesEvent | 2017-05-10 09:32:58 | canonical-websites/www.ubuntu.com | https://api.github.com/repos/canonical-websites/www.ubuntu.com | closed | Top and bottom padding of footer should be the same (L screen) | Review: Design +1 Status: Review | Top seems bigger now. Reviewing /about section | 1.0 | Top and bottom padding of footer should be the same (L screen) - Top seems bigger now. Reviewing /about section | non_test | top and bottom padding of footer should be the same l screen top seems bigger now reviewing about section | 0 |
1,892 | 6,894,596,379 | IssuesEvent | 2017-11-23 10:35:21 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | fabric8.io pipelines stability | area/architecture area/pipelines kind/bug SEV1-urgent team/build-cd team/service-delivery | The build pipelines, running on fabric8.io, go down too often and do not have enough support to keep them running for devs around the world. | 1.0 | fabric8.io pipelines stability - The build pipelines, running on fabric8.io, go down too often and do not have enough support to keep them running for devs around the world. | non_test | io pipelines stability the build pipelines running on io go down too often and do not have enough support to keep them running for devs around the world | 0 |
128,293 | 10,523,964,011 | IssuesEvent | 2019-09-30 12:18:20 | zio/zio | https://api.github.com/repos/zio/zio | closed | Flaky test in FiberRefSpec | tests | See https://circleci.com/gh/zio/zio/28535
```
- the value of the looser is inherited in zipPar
[info] update1 did not satisfy equalTo(update2)
```
cc @zezutom | 1.0 | Flaky test in FiberRefSpec - See https://circleci.com/gh/zio/zio/28535
```
- the value of the looser is inherited in zipPar
[info] update1 did not satisfy equalTo(update2)
```
cc @zezutom | test | flaky test in fiberrefspec see the value of the looser is inherited in zippar did not satisfy equalto cc zezutom | 1 |
494,957 | 14,269,235,977 | IssuesEvent | 2020-11-21 00:39:46 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | opened | [dv] investigate ALL #delay waits with DV testbench | Component:DV Priority:P3 Type:Enhancement | From PR #4282 Sri recommended to investigate all #delay waits and see if we can replace with something better.
I will start to look into them from common DV functions to each IP. | 1.0 | [dv] investigate ALL #delay waits with DV testbench - From PR #4282 Sri recommended to investigate all #delay waits and see if we can replace with something better.
I will start to look into them from common DV functions to each IP. | non_test | investigate all delay waits with dv testbench from pr sri recommended to investigate all delay waits and see if we can replace with something better i will start to look into them from common dv functions to each ip | 0 |
228,709 | 17,470,200,660 | IssuesEvent | 2021-08-07 01:58:48 | fastjengine/FastJ | https://api.github.com/repos/fastjengine/FastJ | closed | [Documentation]: Add Discord server badge & link to FastJ | documentation | ### The Outdated/Incorrect Information
The lack of discord server link is quite bad, let's be honest.
### Suggested Documentation Changes
Add the discord server badge (from shields.io) and a mention to the discord server in the main README.
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | 1.0 | [Documentation]: Add Discord server badge & link to FastJ - ### The Outdated/Incorrect Information
The lack of discord server link is quite bad, let's be honest.
### Suggested Documentation Changes
Add the discord server badge (from shields.io) and a mention to the discord server in the main README.
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | non_test | add discord server badge link to fastj the outdated incorrect information the lack of discord server link is quite bad let s be honest suggested documentation changes add the discord server badge from shields io and a mention to the discord server in the main readme code of conduct i agree to follow this project s code of conduct | 0 |
5,738 | 7,315,822,628 | IssuesEvent | 2018-03-01 12:31:38 | Azure/azure-powershell | https://api.github.com/repos/Azure/azure-powershell | closed | Get-AzureRmRecoveryServicesBackupContainer Parameter -FriendlyName not working | RecoveryServices.Backup | ### Cmdlet(s)
Get-AzureRmRecoveryServicesBackupContainer
### PowerShell Version
Name Value
---- -----
PSVersion 5.1.16299.98
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.16299.98
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
### Module Version
### OS Version
Instructions: to get OS version, type `$PSversionTable` and look for value associated with `BuildVersion`
### Description
Get-AzureRmRecoveryServicesBackupContainer should be used like: Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName $VMname. But the Parameter -FriendlyName does not work at all, no output when using a variable as friendly name or if I enter the VMname String directly
### Debug Output
Instructions: to get Debug Output, set `$DebugPreference="Continue"` and then execute the cmdlet or script causing the issue
### Script/Steps for Reproduction
$vmname = "backuptestvm"
$vmrg = "VMResourceGroup"
$rsv = Get-AzureRmRecoveryServicesVault -ResourceGroupName $VMRg
Set-AzureRmRecoveryServicesVaultContext -Vault $rsv
#with Variable:
$BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName $VMname
#OR with String:
$BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName 'testbackupvm'
#BUT: query without -FriendlyName returns all existing containers in that vault
#BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" | 1.0 | Get-AzureRmRecoveryServicesBackupContainer Parameter -FriendlyName not working - ### Cmdlet(s)
Get-AzureRmRecoveryServicesBackupContainer
### PowerShell Version
Name Value
---- -----
PSVersion 5.1.16299.98
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.16299.98
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
### Module Version
### OS Version
Instructions: to get OS version, type `$PSversionTable` and look for value associated with `BuildVersion`
### Description
Get-AzureRmRecoveryServicesBackupContainer should be used like: Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName $VMname. But the Parameter -FriendlyName does not work at all, no output when using a variable as friendly name or if I enter the VMname String directly
### Debug Output
Instructions: to get Debug Output, set `$DebugPreference="Continue"` and then execute the cmdlet or script causing the issue
### Script/Steps for Reproduction
$vmname = "backuptestvm"
$vmrg = "VMResourceGroup"
$rsv = Get-AzureRmRecoveryServicesVault -ResourceGroupName $VMRg
Set-AzureRmRecoveryServicesVaultContext -Vault $rsv
#with Variable:
$BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName $VMname
#OR with String:
$BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" -FriendlyName 'testbackupvm'
#BUT: query without -FriendlyName returns all existing containers in that vault
#BackupContainer = Get-AzureRmRecoveryServicesBackupContainer -ContainerType "AzureVM" –Status "Registered" | non_test | get azurermrecoveryservicesbackupcontainer parameter friendlyname not working cmdlet s get azurermrecoveryservicesbackupcontainer powershell version name value psversion psedition desktop pscompatibleversions buildversion clrversion wsmanstackversion psremotingprotocolversion serializationversion module version os version instructions to get os version type psversiontable and look for value associated with buildversion description get azurermrecoveryservicesbackupcontainer should be used like get azurermrecoveryservicesbackupcontainer containertype azurevm –status registered friendlyname vmname but the parameter friendlyname does not work at all no output when using a variable as friendly name or if i enter the vmname string directly debug output instructions to get debug output set debugpreference continue and then execute the cmdlet or script causing the issue script steps for reproduction vmname backuptestvm vmrg vmresourcegroup rsv get azurermrecoveryservicesvault resourcegroupname vmrg set azurermrecoveryservicesvaultcontext vault rsv with variable backupcontainer get azurermrecoveryservicesbackupcontainer containertype azurevm –status registered friendlyname vmname or with string backupcontainer get azurermrecoveryservicesbackupcontainer containertype azurevm –status registered friendlyname testbackupvm but query without friendlyname returns all existing containers in that vault backupcontainer get azurermrecoveryservicesbackupcontainer containertype azurevm –status registered | 0 |
22,349 | 11,579,088,970 | IssuesEvent | 2020-02-21 17:10:13 | typelevel/kittens | https://api.github.com/repos/typelevel/kittens | closed | Performance Regression | Performance | When updating version numbers for [scalaz-and-cats](https://github.com/fosskers/scalaz-and-cats) I noticed a fairly stark performance regression for kittens-derived `Eq` between RC1 and RC3.
For a simple case-class:
```scala
case class Foo(age: Int, msg: String, truthy: Boolean)
```
Previously the deriving syntax was:
```scala
implicit val fooEq: Eq[Foo] = derive.eq[Foo]
```
And now it is:
```scala
implicit val fooEq: Eq[Foo] = {
import derived.auto.eq._
derived.semi.eq
}
```
For two length-10000 `List[Foo]` who have the same contents but are not the same object, to compare every element dropped from **38,630 ns** to **45,416 ns**. Either number is an order of magnitude slower than a hand-written instance, vanilla Scala's `==`, or a ScalaZ auto-derived instance. | True | Performance Regression - When updating version numbers for [scalaz-and-cats](https://github.com/fosskers/scalaz-and-cats) I noticed a fairly stark performance regression for kittens-derived `Eq` between RC1 and RC3.
For a simple case-class:
```scala
case class Foo(age: Int, msg: String, truthy: Boolean)
```
Previously the deriving syntax was:
```scala
implicit val fooEq: Eq[Foo] = derive.eq[Foo]
```
And now it is:
```scala
implicit val fooEq: Eq[Foo] = {
import derived.auto.eq._
derived.semi.eq
}
```
For two length-10000 `List[Foo]` who have the same contents but are not the same object, to compare every element dropped from **38,630 ns** to **45,416 ns**. Either number is an order of magnitude slower than a hand-written instance, vanilla Scala's `==`, or a ScalaZ auto-derived instance. | non_test | performance regression when updating version numbers for i noticed a fairly stark performance regression for kittens derived eq between and for a simple case class scala case class foo age int msg string truthy boolean previously the deriving syntax was scala implicit val fooeq eq derive eq and now it is scala implicit val fooeq eq import derived auto eq derived semi eq for two length list who have the same contents but are not the same object to compare every element dropped from ns to ns either number is an order of magnitude slower than a hand written instance vanilla scala s or a scalaz auto derived instance | 0 |
1,306 | 2,933,750,222 | IssuesEvent | 2015-06-30 01:45:06 | OmniLayer/omnicore | https://api.github.com/repos/OmniLayer/omnicore | closed | Hold LOCKs only as long as necessary | improvement performance | As outlined in #84,`gettransactions_MP` and other parts currently lock certain sections longer than necessary.
In case of `gettransactions_MP` an additional improvement could be to restructure the command, such that we lock the wallet, collect the wallet transactions, release the lock, and then process the transactions, instead of holding the lock for the whole time.
I'm going to tackle it, once the RPC branch is in. | True | Hold LOCKs only as long as necessary - As outlined in #84,`gettransactions_MP` and other parts currently lock certain sections longer than necessary.
In case of `gettransactions_MP` an additional improvement could be to restructure the command, such that we lock the wallet, collect the wallet transactions, release the lock, and then process the transactions, instead of holding the lock for the whole time.
I'm going to tackle it, once the RPC branch is in. | non_test | hold locks only as long as necessary as outlined in gettransactions mp and other parts currently lock certain sections longer than necessary in case of gettransactions mp an additional improvement could be to restructure the command such that we lock the wallet collect the wallet transactions release the lock and then process the transactions instead of holding the lock for the whole time i m going to tackle it once the rpc branch is in | 0 |
622,458 | 19,636,042,077 | IssuesEvent | 2022-01-08 09:44:43 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | opened | Light Religion Changes | suggestion :question: priority low :grey_exclamation: | <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
I would like to propose a few changes in the Light faiths.
The first one would be adding Stratholme as one of the Holy Light holy sites, since it is one in the lore according to https://wowpedia.fandom.com/wiki/Stratholme.
To quote the link, "Among the oldest and most revered of these holy sites were the temples in the Eastweald, including Stratholme." and "It was also a holy site, being the Silver Hand paladin order's birthplace.".
I think CK3 does allow for more than 5 holy sites, but if you really need to swap it with another holy site, I would suggest swapping it with the holy site of Ironforge, as I can't really see why it is a holy site in the first place, since the dwarves now have their own Light faith, Khazism, making the main Holy Light faith more human centered.
I can only assume that's a leftover from the CK2 mod where the dwarves and the humans had the same religion.
The other change I'd like to propose would be to make order or arcane magic accepted in the Scarlet Light faith, since they do have mages in their order, the most notable being https://wowpedia.fandom.com/wiki/Arcanist_Doan and https://wowpedia.fandom.com/wiki/Flameweaver_Koegler.
In fact I'd even go as far as to have them accept spirit magic, or whichever doctrine will allow for the monk class, as I remember there being monks in the Scarlet Monastery dungeon(https://wowpedia.fandom.com/wiki/Scarlet_Monk).
My final suggestion would be the implementation of an Argent Dawn faith.
I already know that this faith is planned, but I'd like to put some input in the design.
As they accept Undead into their ranks, they should have a doctrine similar to the Tauren faith Earth Mother's Mercy, where being undead isn't a crime in and of itself.
As for the tenets, I'd suggest Armed Pilgrimages, since they are a crusade, Adaptive, since they even accept members of Horde races into their ranks, and maybe Unrelenting Faith, since they are militant.
As for the holy sites they'd be the same as the ones of Holy Light plus Icecrown. | 1.0 | Light Religion Changes - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
I would like to propose a few changes in the Light faiths.
The first one would be adding Stratholme as one of the Holy Light holy sites, since it is one in the lore according to https://wowpedia.fandom.com/wiki/Stratholme.
To quote the link, "Among the oldest and most revered of these holy sites were the temples in the Eastweald, including Stratholme." and "It was also a holy site, being the Silver Hand paladin order's birthplace.".
I think CK3 does allow for more than 5 holy sites, but if you really need to swap it with another holy site, I would suggest swapping it with the holy site of Ironforge, as I can't really see why it is a holy site in the first place, since the dwarves now have their own Light faith, Khazism, making the main Holy Light faith more human centered.
I can only assume that's a leftover from the CK2 mod where the dwarves and the humans had the same religion.
The other change I'd like to propose would be to make order or arcane magic accepted in the Scarlet Light faith, since they do have mages in their order, the most notable being https://wowpedia.fandom.com/wiki/Arcanist_Doan and https://wowpedia.fandom.com/wiki/Flameweaver_Koegler.
In fact I'd even go as far as to have them accept spirit magic, or whichever doctrine will allow for the monk class, as I remember there being monks in the Scarlet Monastery dungeon(https://wowpedia.fandom.com/wiki/Scarlet_Monk).
My final suggestion would be the implementation of an Argent Dawn faith.
I already know that this faith is planned, but I'd like to put some input in the design.
As they accept Undead into their ranks, they should have a doctrine similar to the Tauren faith Earth Mother's Mercy, where being undead isn't a crime in and of itself.
As for the tenets, I'd suggest Armed Pilgrimages, since they are a crusade, Adaptive, since they even accept members of Horde races into their ranks, and maybe Unrelenting Faith, since they are militant.
As for the holy sites they'd be the same as the ones of Holy Light plus Icecrown. | non_test | light religion changes do not remove pre existing lines describe your suggestion in full detail below i would like to propose a few changes in the light faiths the first one would be adding stratholme as one of the holy light holy sites since it is one in the lore according to to quote the link among the oldest and most revered of these holy sites were the temples in the eastweald including stratholme and it was also a holy site being the silver hand paladin order s birthplace i think does allow for more than holy sites but if you really need to swap it with another holy site i would suggest swapping it with the holy site of ironforge as i can t really see why it is a holy site in the first place since the dwarves now have their own light faith khazism making the main holy light faith more human centered i can only assume that s a leftover from the mod where the dwarves and the humans had the same religion the other change i d like to propose would be to make order or arcane magic accepted in the scarlet light faith since they do have mages in their order the most notable being and in fact i d even go as far as to have them accept spirit magic or whichever doctrine will allow for the monk class as i remember there being monks in the scarlet monastery dungeon my final suggestion would be the implementation of an argent dawn faith i already know that this faith is planned but i d like to put some input in the design as they accept undead into their ranks they should have a doctrine similar to the tauren faith earth mother s mercy where being undead isn t a crime in and of itself as for the tenets i d suggest armed pilgrimages since they are a crusade adaptive since they even accept members of horde races into their ranks and maybe unrelenting faith since they are militant as for the holy sites they d be the same as the ones of holy light plus icecrown | 0 |
99,184 | 4,049,127,056 | IssuesEvent | 2016-05-23 13:09:51 | duckduckgo/community-platform | https://api.github.com/repos/duckduckgo/community-platform | closed | insert image does not allow upload | Feature Forum Improvement Priority: Low | current forum post `insert image` button prompts for an image url but does not allow a new image to be uploaded to dukgo.com | 1.0 | insert image does not allow upload - current forum post `insert image` button prompts for an image url but does not allow a new image to be uploaded to dukgo.com | non_test | insert image does not allow upload current forum post insert image button prompts for an image url but does not allow a new image to be uploaded to dukgo com | 0 |
198,473 | 14,981,777,393 | IssuesEvent | 2021-01-28 15:14:01 | zeebe-io/zeebe | https://api.github.com/repos/zeebe-io/zeebe | opened | WorkflowExecutionCleanStateTest#testWorkflowWithEventSubprocess | Type: Unstable Test | **Summary**
> How often does the test fail?
Seldom (at least I haven't seen it before)
> Does it block your work?
No.
> Do we suspect that it is a real failure?
Based on the assertion it might simply be a condition that took too long to be met but I didn't look into it.
**Failures**
> Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins
<details><summary>Assertion failure</summary>
<pre>
org.awaitility.core.ConditionTimeoutException:
Assertion condition defined as a lambda expression in io.zeebe.engine.state.WorkflowExecutionCleanStateTest [Expected all columns to be empty]
Expecting empty but was: [JOBS, JOB_STATES, JOB_ACTIVATABLE] within 10 seconds.
at org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:165)
at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:119)
at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:31)
at org.awaitility.core.ConditionFactory.until(ConditionFactory.java:895)
at org.awaitility.core.ConditionFactory.untilAsserted(ConditionFactory.java:679)
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.assertThatStateIsEmpty(WorkflowExecutionCleanStateTest.java:601)
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.testWorkflowWithEventSubprocess(WorkflowExecutionCleanStateTest.java:507)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:43)
at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:82)
at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:73)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:108)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:96)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:75)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.lambda$execute$1(JUnitPlatformProvider.java:199)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:193)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:120)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
Caused by: java.lang.AssertionError: [Expected all columns to be empty]
Expecting empty but was: [JOBS, JOB_STATES, JOB_ACTIVATABLE]
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.lambda$assertThatStateIsEmpty$18(WorkflowExecutionCleanStateTest.java:609)
at org.awaitility.core.AssertionCondition.lambda$new$0(AssertionCondition.java:53)
at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:222)
at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:209)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)</pre>
</details>
**Logs**
https://ci.zeebe.camunda.cloud/blue/organizations/jenkins/zeebe-io%2Fzeebe/detail/staging/2780/tests
<details><summary>Logs</summary>
<pre>
13:39:01.052 [Broker-0-LogStream-1] DEBUG io.zeebe.logstreams - Configured log appender back pressure at partition 1 as AppenderVegasCfg{initialLimit=1024, maxConcurrency=32768, alphaLimit=0.7, betaLimit=0.95}. Window limiting is disabled
13:39:01.346 [Broker-0-StreamProcessor-1] DEBUG io.zeebe.logstreams - Recovering state of partition 1 from snapshot
13:39:01.347 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@72c9311f)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@7c0ed136, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@78315a6f, configuration: Configuration(false)]
13:39:01.350 [Broker-0-StreamProcessor-1] INFO io.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1
13:39:01.351 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@3d33db0d)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@6c9547d, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@c44a2f3, configuration: Configuration(false)]
13:39:01.353 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@7e547336)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@5d1c2442, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@11bae920, configuration: Configuration(false)]
13:39:01.586 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@7b3b0772)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@5e7965c4, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@478f6de5, configuration: Configuration(false)]
13:39:12.059 [] INFO io.zeebe.test.records - Test failed, following records were exported:
13:39:12.258 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":-1,"position":1,"timestamp":1611841141543,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":-1,"value":{"deployedWorkflows":[],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlb...
13:39:12.262 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":2,"timestamp":1611841141685,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":1,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwM...
13:39:12.264 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":3,"timestamp":1611841141738,"recordType":"COMMAND","intent":"DISTRIBUTE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":2,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1O...
13:39:12.265 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":4,"timestamp":1611841141743,"recordType":"EVENT","intent":"DISTRIBUTED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":3,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OL...
13:39:12.348 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE_CREATION","key":-1,"position":5,"timestamp":1611841141745,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":-1,"value":{"variables":{},"version":-1,"workflowInstanceKey":-1,"bpmnProcessId":"workflow","workflowKey":-1}}
13:39:12.356 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":6,"timestamp":1611841141747,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":5,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.357 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE_CREATION","key":2251799813685252,"position":7,"timestamp":1611841141747,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":5,"value":{"variables":{},"version":1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249}}
13:39:12.362 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":-1,"position":8,"timestamp":1611841141753,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":6,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.363 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":9,"timestamp":1611841141753,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":6,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.364 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":10,"timestamp":1611841141755,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":8,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.365 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":11,"timestamp":1611841141758,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":9,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.365 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":12,"timestamp":1611841141763,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":11,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.366 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":13,"timestamp":1611841141764,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":12,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.366 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":14,"timestamp":1611841141767,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":13,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.367 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685255,"position":15,"timestamp":1611841141771,"recordType":"EVENT","intent":"SEQUENCE_FLOW_TAKEN","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":14,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"sequenceFlow_2cb253c1-5f5f-4fbc-848c-17b928237982","bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.374 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":16,"timestamp":1611841141854,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":15,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.374 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":17,"timestamp":1611841141854,"recordType":"COMMAND","intent":"TRIGGER","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":15,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.375 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":18,"timestamp":1611841141858,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":16,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.375 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685258,"position":19,"timestamp":1611841141863,"recordType":"EVENT","intent":"EVENT_OCCURRED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":17,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.376 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":20,"timestamp":1611841141863,"recordType":"EVENT","intent":"TRIGGERED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":17,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.384 [] INFO io.zeebe.test.records - {"valueType":"JOB","key":-1,"position":21,"timestamp":1611841141864,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":18,"value":{"deadline":-1,"variables":{},"errorMessage":"","type":"test","workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","elementInstanceKey":2251799813685256,"retries":3,"customHeaders":{},"worker":"","workflowDefinitionVersion":1,"errorCode":""}}
13:39:12.385 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":22,"timestamp":1611841141870,"recordType":"EVENT","intent":"ELEMENT_TERMINATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":19,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.385 [] INFO io.zeebe.test.records - {"valueType":"JOB","key":2251799813685260,"position":23,"timestamp":1611841141872,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":21,"value":{"deadline":-1,"variables":{},"errorMessage":"","type":"test","workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","elementInstanceKey":2251799813685256,"retries":3,"customHeaders":{},"worker":"","workflowDefinitionVersion":1,"errorCode":""}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":24,"timestamp":1611841141879,"recordType":"EVENT","intent":"ELEMENT_TERMINATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":22,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":25,"timestamp":1611841141937,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":24,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":26,"timestamp":1611841141940,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":25,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.387 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":27,"timestamp":1611841141942,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":26,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.391 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":28,"timestamp":1611841141944,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":27,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.391 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":29,"timestamp":1611841141947,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":28,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.392 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":30,"timestamp":1611841141948,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":29,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.392 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685262,"position":31,"timestamp":1611841141950,"recordType":"EVENT","intent":"SEQUENCE_FLOW_TAKEN","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":30,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"sequenceFlow_878d5d38-89fa-4ba9-9a60-e8f2199032da","bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.393 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":32,"timestamp":1611841141952,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":31,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.393 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":33,"timestamp":1611841141953,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":32,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.394 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":34,"timestamp":1611841141954,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":33,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.395 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":35,"timestamp":1611841141955,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":34,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.437 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":36,"timestamp":1611841141957,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":35,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.437 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":37,"timestamp":1611841141959,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":36,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.438 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":38,"timestamp":1611841141961,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":37,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.438 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":39,"timestamp":1611841141963,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":38,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.441 [] INFO io.zeebe.broker.test - Test failed, following records were exported:
13:39:12.481 [] INFO io.zeebe.broker.test - Records on partition 1:
Position: 1 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[]}
Position: 2 Key: 2251799813685250
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 3 Key: 2251799813685250
RecordMetadata{recordType=COMMAND, intentValue=255, intent=DISTRIBUTE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 4 Key: 2251799813685250
RecordMetadata{recordType=EVENT, intentValue=255, intent=DISTRIBUTED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 5 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE_CREATION, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","workflowKey":-1,"workflowInstanceKey":-1,"version":-1,"variables":"gA==","fetchVariables":[]}
Position: 6 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 7 Key: 2251799813685252
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE_CREATION, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"version":1,"variables":"gA==","fetchVariables":[]}
Position: 8 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 9 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 10 Key: 2251799813685253
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 11 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 12 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 13 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 14 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 15 Key: 2251799813685255
RecordMetadata{recordType=EVENT, intentValue=255, intent=SEQUENCE_FLOW_TAKEN, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"sequenceFlow_2cb253c1-5f5f-4fbc-848c-17b928237982","flowScopeKey":2251799813685251,"bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 16 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 17 Key: 2251799813685253
RecordMetadata{recordType=COMMAND, intentValue=255, intent=TRIGGER, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 18 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 19 Key: 2251799813685258
RecordMetadata{recordType=EVENT, intentValue=255, intent=EVENT_OCCURRED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 20 Key: 2251799813685253
RecordMetadata{recordType=EVENT, intentValue=255, intent=TRIGGERED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 21 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=JOB, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"deadline":-1,"worker":"","retries":3,"type":"test","customHeaders":[packed value (length=1)],"variables":"gA==","errorMessage":"","errorCode":"","bpmnProcessId":"workflow","workflowDefinitionVersion":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","elementInstanceKey":2251799813685256}
Position: 22 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_TERMINATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 23 Key: 2251799813685260
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=JOB, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"deadline":-1,"worker":"","retries":3,"type":"test","customHeaders":[packed value (length=1)],"variables":"gA==","errorMessage":"","errorCode":"","bpmnProcessId":"workflow","workflowDefinitionVersion":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","elementInstanceKey":2251799813685256}
Position: 24 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_TERMINATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 25 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 26 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 27 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 28 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 29 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 30 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 31 Key: 2251799813685262
RecordMetadata{recordType=EVENT, intentValue=255, intent=SEQUENCE_FLOW_TAKEN, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"sequenceFlow_878d5d38-89fa-4ba9-9a60-e8f2199032da","flowScopeKey":2251799813685259,"bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 32 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 33 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 34 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 35 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 36 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 37 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 38 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 39 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
13:39:12.489 [] DEBUG io.zeebe.util.buffer - Close stream processor
13:39:12.490 [Broker-0-StreamProcessor-1] DEBUG io.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1.
13:39:12.588 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - Close appender for log stream stream-1
13:39:12.589 [stream-1-write-buffer] DEBUG io.zeebe.dispatcher - Dispatcher closed
13:39:12.590 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - On closing logstream stream-1 close 4 readers
13:39:12.590 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - Close log storage with name stream-1
13:39:12.637 [] DEBUG io.zeebe.broker.test - Clean up test files on path /tmp/junit17217163829394622494
13:39:12.638 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers'
13:39:12.638 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-actors'
13:39:12.639 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers': closed successfully
13:39:12.640 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-actors': closed successfully</pre>
</details>
| 1.0 | WorkflowExecutionCleanStateTest#testWorkflowWithEventSubprocess - **Summary**
> How often does the test fail?
Seldom (at least I haven't seen it before)
> Does it block your work?
No.
> Do we suspect that it is a real failure?
Based on the assertion it might simply be a condition that took too long to be met but I didn't look into it.
**Failures**
> Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins
<details><summary>Assertion failure</summary>
<pre>
org.awaitility.core.ConditionTimeoutException:
Assertion condition defined as a lambda expression in io.zeebe.engine.state.WorkflowExecutionCleanStateTest [Expected all columns to be empty]
Expecting empty but was: [JOBS, JOB_STATES, JOB_ACTIVATABLE] within 10 seconds.
at org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:165)
at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:119)
at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:31)
at org.awaitility.core.ConditionFactory.until(ConditionFactory.java:895)
at org.awaitility.core.ConditionFactory.untilAsserted(ConditionFactory.java:679)
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.assertThatStateIsEmpty(WorkflowExecutionCleanStateTest.java:601)
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.testWorkflowWithEventSubprocess(WorkflowExecutionCleanStateTest.java:507)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:43)
at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at java.base/java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:82)
at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:73)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:108)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:96)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:75)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.lambda$execute$1(JUnitPlatformProvider.java:199)
at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:193)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:120)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
Caused by: java.lang.AssertionError: [Expected all columns to be empty]
Expecting empty but was: [JOBS, JOB_STATES, JOB_ACTIVATABLE]
at io.zeebe.engine.state.WorkflowExecutionCleanStateTest.lambda$assertThatStateIsEmpty$18(WorkflowExecutionCleanStateTest.java:609)
at org.awaitility.core.AssertionCondition.lambda$new$0(AssertionCondition.java:53)
at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:222)
at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:209)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)</pre>
</details>
**Logs**
https://ci.zeebe.camunda.cloud/blue/organizations/jenkins/zeebe-io%2Fzeebe/detail/staging/2780/tests
<details><summary>Logs</summary>
<pre>
13:39:01.052 [Broker-0-LogStream-1] DEBUG io.zeebe.logstreams - Configured log appender back pressure at partition 1 as AppenderVegasCfg{initialLimit=1024, maxConcurrency=32768, alphaLimit=0.7, betaLimit=0.95}. Window limiting is disabled
13:39:01.346 [Broker-0-StreamProcessor-1] DEBUG io.zeebe.logstreams - Recovering state of partition 1 from snapshot
13:39:01.347 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@72c9311f)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@7c0ed136, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@78315a6f, configuration: Configuration(false)]
13:39:01.350 [Broker-0-StreamProcessor-1] INFO io.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1
13:39:01.351 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@3d33db0d)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@6c9547d, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@c44a2f3, configuration: Configuration(false)]
13:39:01.353 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@7e547336)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@5d1c2442, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@11bae920, configuration: Configuration(false)]
13:39:01.586 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.zeebe.el.impl.feel.MessagePackValueMapper@7b3b0772)), function-provider: io.zeebe.el.impl.feel.FeelFunctionProvider@5e7965c4, clock: io.zeebe.el.impl.ZeebeFeelEngineClock@478f6de5, configuration: Configuration(false)]
13:39:12.059 [] INFO io.zeebe.test.records - Test failed, following records were exported:
13:39:12.258 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":-1,"position":1,"timestamp":1611841141543,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":-1,"value":{"deployedWorkflows":[],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlb...
13:39:12.262 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":2,"timestamp":1611841141685,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":1,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwM...
13:39:12.264 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":3,"timestamp":1611841141738,"recordType":"COMMAND","intent":"DISTRIBUTE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":2,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1O...
13:39:12.265 [] INFO io.zeebe.test.records - {"valueType":"DEPLOYMENT","key":2251799813685250,"position":4,"timestamp":1611841141743,"recordType":"EVENT","intent":"DISTRIBUTED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":3,"value":{"deployedWorkflows":[{"version":1,"resourceName":"process.xml","bpmnProcessId":"workflow","workflowKey":2251799813685249}],"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OL...
13:39:12.348 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE_CREATION","key":-1,"position":5,"timestamp":1611841141745,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":-1,"value":{"variables":{},"version":-1,"workflowInstanceKey":-1,"bpmnProcessId":"workflow","workflowKey":-1}}
13:39:12.356 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":6,"timestamp":1611841141747,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":5,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.357 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE_CREATION","key":2251799813685252,"position":7,"timestamp":1611841141747,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":5,"value":{"variables":{},"version":1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249}}
13:39:12.362 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":-1,"position":8,"timestamp":1611841141753,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":6,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.363 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":9,"timestamp":1611841141753,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":6,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.364 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":10,"timestamp":1611841141755,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":8,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.365 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":11,"timestamp":1611841141758,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":9,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.365 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":12,"timestamp":1611841141763,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":11,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.366 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":13,"timestamp":1611841141764,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":12,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.366 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685254,"position":14,"timestamp":1611841141767,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":13,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.367 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685255,"position":15,"timestamp":1611841141771,"recordType":"EVENT","intent":"SEQUENCE_FLOW_TAKEN","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":14,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"sequenceFlow_2cb253c1-5f5f-4fbc-848c-17b928237982","bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.374 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":16,"timestamp":1611841141854,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":15,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.374 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":17,"timestamp":1611841141854,"recordType":"COMMAND","intent":"TRIGGER","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":15,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.375 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":18,"timestamp":1611841141858,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":16,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.375 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685258,"position":19,"timestamp":1611841141863,"recordType":"EVENT","intent":"EVENT_OCCURRED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":17,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.376 [] INFO io.zeebe.test.records - {"valueType":"TIMER","key":2251799813685253,"position":20,"timestamp":1611841141863,"recordType":"EVENT","intent":"TRIGGERED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":17,"value":{"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","workflowInstanceKey":2251799813685251,"workflowKey":2251799813685249,"repetitions":1,"dueDate":1611841141852,"elementInstanceKey":2251799813685251}}
13:39:12.384 [] INFO io.zeebe.test.records - {"valueType":"JOB","key":-1,"position":21,"timestamp":1611841141864,"recordType":"COMMAND","intent":"CREATE","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":18,"value":{"deadline":-1,"variables":{},"errorMessage":"","type":"test","workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","elementInstanceKey":2251799813685256,"retries":3,"customHeaders":{},"worker":"","workflowDefinitionVersion":1,"errorCode":""}}
13:39:12.385 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":22,"timestamp":1611841141870,"recordType":"EVENT","intent":"ELEMENT_TERMINATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":19,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.385 [] INFO io.zeebe.test.records - {"valueType":"JOB","key":2251799813685260,"position":23,"timestamp":1611841141872,"recordType":"EVENT","intent":"CREATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":21,"value":{"deadline":-1,"variables":{},"errorMessage":"","type":"test","workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","elementInstanceKey":2251799813685256,"retries":3,"customHeaders":{},"worker":"","workflowDefinitionVersion":1,"errorCode":""}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685256,"position":24,"timestamp":1611841141879,"recordType":"EVENT","intent":"ELEMENT_TERMINATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":22,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"task","bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":25,"timestamp":1611841141937,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":24,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.386 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":26,"timestamp":1611841141940,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":25,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.387 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":27,"timestamp":1611841141942,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":26,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.391 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":28,"timestamp":1611841141944,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":27,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.391 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":29,"timestamp":1611841141947,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":28,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.392 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685261,"position":30,"timestamp":1611841141948,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":29,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.392 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685262,"position":31,"timestamp":1611841141950,"recordType":"EVENT","intent":"SEQUENCE_FLOW_TAKEN","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":30,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"sequenceFlow_878d5d38-89fa-4ba9-9a60-e8f2199032da","bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.393 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":32,"timestamp":1611841141952,"recordType":"EVENT","intent":"ELEMENT_ACTIVATING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":31,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.393 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":33,"timestamp":1611841141953,"recordType":"EVENT","intent":"ELEMENT_ACTIVATED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":32,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.394 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":34,"timestamp":1611841141954,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":33,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.395 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685263,"position":35,"timestamp":1611841141955,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":34,"value":{"version":1,"flowScopeKey":2251799813685259,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.437 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":36,"timestamp":1611841141957,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":35,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.437 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685259,"position":37,"timestamp":1611841141959,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":36,"value":{"version":1,"flowScopeKey":2251799813685251,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"event-subprocess","bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.438 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":38,"timestamp":1611841141961,"recordType":"EVENT","intent":"ELEMENT_COMPLETING","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":37,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.438 [] INFO io.zeebe.test.records - {"valueType":"WORKFLOW_INSTANCE","key":2251799813685251,"position":39,"timestamp":1611841141963,"recordType":"EVENT","intent":"ELEMENT_COMPLETED","partitionId":1,"rejectionType":"NULL_VAL","rejectionReason":"","brokerVersion":"0.27.0","sourceRecordPosition":38,"value":{"version":1,"flowScopeKey":-1,"workflowInstanceKey":2251799813685251,"bpmnProcessId":"workflow","workflowKey":2251799813685249,"elementId":"workflow","bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}}
13:39:12.441 [] INFO io.zeebe.broker.test - Test failed, following records were exported:
13:39:12.481 [] INFO io.zeebe.broker.test - Records on partition 1:
Position: 1 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[]}
Position: 2 Key: 2251799813685250
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 3 Key: 2251799813685250
RecordMetadata{recordType=COMMAND, intentValue=255, intent=DISTRIBUTE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 4 Key: 2251799813685250
RecordMetadata{recordType=EVENT, intentValue=255, intent=DISTRIBUTED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=DEPLOYMENT, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"resources":[{"resourceType":"BPMN_XML","resourceName":"process.xml","resource":"PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0iVVRGLTgiIHN0YW5kYWxvbmU9Im5vIj8+CjxkZWZpbml0aW9ucyB4bWxuczpicG1uZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvREkiIHhtbG5zOmRjPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9ERC8yMDEwMDUyNC9EQyIgeG1sbnM6ZGk9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0RELzIwMTAwNTI0L0RJIiB4bWxuczpuczA9Imh0dHA6Ly9jYW11bmRhLm9yZy9zY2hlbWEvemVlYmUvMS4wIiBpZD0iZGVmaW5pdGlvbnNfNDlmMDM3NDMtMjcwMy00NTdlLWJiM2QtNTkyNmQ0OWZlM2FjIiB0YXJnZXROYW1lc3BhY2U9Imh0dHA6Ly93d3cub21nLm9yZy9zcGVjL0JQTU4vMjAxMDA1MjQvTU9ERUwiIHhtbG5zPSJodHRwOi8vd3d3Lm9tZy5vcmcvc3BlYy9CUE1OLzIwMTAwNTI0L01PREVMIj4KICA8cHJvY2VzcyBpZD0id29ya2Zsb3ciIGlzRXhlY3V0YWJsZT0idHJ1ZSI+CiAgICA8c3ViUHJvY2VzcyBpZD0iZXZlbnQtc3VicHJvY2VzcyIgbmFtZT0iZXZlbnQtc3VicHJvY2VzcyIgdHJpZ2dlcmVkQnlFdmVudD0idHJ1ZSI+CiAgICAgIDxzdGFydEV2ZW50IGlkPSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaXNJbnRlcnJ1cHRpbmc9InRydWUiPgogICAgICAgIDxvdXRnb2luZz5zZXF1ZW5jZUZsb3dfODc4ZDVkMzgtODlmYS00YmE5LTlhNjAtZThmMjE5OTAzMmRhPC9vdXRnb2luZz4KICAgICAgICA8dGltZXJFdmVudERlZmluaXRpb24gaWQ9InRpbWVyRXZlbnREZWZpbml0aW9uXzYyODZiZWU2LTk4ZmQtNDFhYi05YzJjLTU1OTg1NmFkNTg3NiI+CiAgICAgICAgICA8dGltZUR1cmF0aW9uIGlkPSJ0aW1lRHVyYXRpb25fNGQ5OTY5N2EtOTljNS00ZTQyLTkxZjktZWI0MzMwNjM5OTI5Ij5QVDAuMVM8L3RpbWVEdXJhdGlvbj4KICAgICAgICA8L3RpbWVyRXZlbnREZWZpbml0aW9uPgogICAgICA8L3N0YXJ0RXZlbnQ+CiAgICAgIDxlbmRFdmVudCBpZD0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ij4KICAgICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYTwvaW5jb21pbmc+CiAgICAgIDwvZW5kRXZlbnQ+CiAgICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd184NzhkNWQzOC04OWZhLTRiYTktOWE2MC1lOGYyMTk5MDMyZGEiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF80ZDA5MjJlZC0zY2JjLTQ0NGEtOWY0NS1jZGQ5N2JlMTMyYjEiIHRhcmdldFJlZj0iZW5kRXZlbnRfOWNhYzBiOWMtYzdmNS00ODBmLWJjOTItMjU4YTE0MDVkYTE4Ii8+CiAgICA8L3N1YlByb2Nlc3M+CiAgICA8c3RhcnRFdmVudCBpZD0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiPgogICAgICA8b3V0Z29pbmc+c2VxdWVuY2VGbG93XzJjYjI1M2MxLTVmNWYtNGZiYy04NDhjLTE3YjkyODIzNzk4Mjwvb3V0Z29pbmc+CiAgICA8L3N0YXJ0RXZlbnQ+CiAgICA8c2VydmljZVRhc2sgaWQ9InRhc2siIG5hbWU9InRhc2siPgogICAgICA8ZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgICAgPG5zMDp0YXNrRGVmaW5pdGlvbiB0eXBlPSJ0ZXN0Ii8+CiAgICAgIDwvZXh0ZW5zaW9uRWxlbWVudHM+CiAgICAgIDxpbmNvbWluZz5zZXF1ZW5jZUZsb3dfMmNiMjUzYzEtNWY1Zi00ZmJjLTg0OGMtMTdiOTI4MjM3OTgyPC9pbmNvbWluZz4KICAgICAgPG91dGdvaW5nPnNlcXVlbmNlRmxvd180ZGRiNjRkMC1jNTVkLTRiNjctOGZkMC1lNmRlZjBkZmIwNTM8L291dGdvaW5nPgogICAgPC9zZXJ2aWNlVGFzaz4KICAgIDxzZXF1ZW5jZUZsb3cgaWQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIHNvdXJjZVJlZj0ic3RhcnRFdmVudF9mYjFkZjU0YS03YWFhLTRlNjctOWQzYi05ZjY5N2MzMDhkYjUiIHRhcmdldFJlZj0idGFzayIvPgogICAgPGVuZEV2ZW50IGlkPSJlbmRFdmVudF8zYjM2Y2RkZC1hMmRjLTRhNWMtOTg4MS03MWRlNjg5MTMyMDciPgogICAgICA8aW5jb21pbmc+c2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MzwvaW5jb21pbmc+CiAgICA8L2VuZEV2ZW50PgogICAgPHNlcXVlbmNlRmxvdyBpZD0ic2VxdWVuY2VGbG93XzRkZGI2NGQwLWM1NWQtNGI2Ny04ZmQwLWU2ZGVmMGRmYjA1MyIgc291cmNlUmVmPSJ0YXNrIiB0YXJnZXRSZWY9ImVuZEV2ZW50XzNiMzZjZGRkLWEyZGMtNGE1Yy05ODgxLTcxZGU2ODkxMzIwNyIvPgogIDwvcHJvY2Vzcz4KICA8YnBtbmRpOkJQTU5EaWFncmFtIGlkPSJCUE1ORGlhZ3JhbV9hNmFkNjdlNy1iMDFlLTRlNTMtYjEzOC0xNjkyZjc1YWJlNmUiPgogICAgPGJwbW5kaTpCUE1OUGxhbmUgYnBtbkVsZW1lbnQ9IndvcmtmbG93IiBpZD0iQlBNTlBsYW5lX2YzZTRhNGJkLWQyMGMtNDNiMi05NDE5LTFhYTQ2NmQwZjYzYiI+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJldmVudC1zdWJwcm9jZXNzIiBpZD0iQlBNTlNoYXBlX2RjODcyNDM3LTg1NTctNGNlYy1hODE1LWY3ODRmYjk1MmM0NiIgaXNFeHBhbmRlZD0idHJ1ZSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjIwMC4wIiB3aWR0aD0iMzUwLjAiIHg9IjEwMC4wIiB5PSIyNTAuMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50XzRkMDkyMmVkLTNjYmMtNDQ0YS05ZjQ1LWNkZDk3YmUxMzJiMSIgaWQ9IkJQTU5TaGFwZV9mZjQ1OTIyYS1hYzkyLTRlZjAtOGEyMi02NWJjMzA2MmZmMDIiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTUwLjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9ImVuZEV2ZW50XzljYWMwYjljLWM3ZjUtNDgwZi1iYzkyLTI1OGExNDA1ZGExOCIgaWQ9IkJQTU5TaGFwZV9jYTg0NzhkOS1hODQ0LTQwNTAtYjJiNC1kOGEwYTNhNzVlMGUiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMjM2LjAiIHk9IjMzMi4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1ORWRnZSBicG1uRWxlbWVudD0ic2VxdWVuY2VGbG93Xzg3OGQ1ZDM4LTg5ZmEtNGJhOS05YTYwLWU4ZjIxOTkwMzJkYSIgaWQ9IkJQTU5FZGdlXzIzOTY0ZjBhLTRiOTktNDg5My1hNjhmLWFmNGY4YzY1NmU4NCI+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjE4Ni4wIiB5PSIzNTAuMCIvPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIyMzYuMCIgeT0iMzUwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTkVkZ2U+CiAgICAgIDxicG1uZGk6QlBNTlNoYXBlIGJwbW5FbGVtZW50PSJzdGFydEV2ZW50X2ZiMWRmNTRhLTdhYWEtNGU2Ny05ZDNiLTlmNjk3YzMwOGRiNSIgaWQ9IkJQTU5TaGFwZV85NzY4MjQ5OC00ZWU2LTQ0NjQtYjM1Ny1iODQ4NzY1ZDQzMzAiPgogICAgICAgIDxkYzpCb3VuZHMgaGVpZ2h0PSIzNi4wIiB3aWR0aD0iMzYuMCIgeD0iMTAwLjAiIHk9IjEwMC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5TaGFwZT4KICAgICAgPGJwbW5kaTpCUE1OU2hhcGUgYnBtbkVsZW1lbnQ9InRhc2siIGlkPSJCUE1OU2hhcGVfMTA5NWFkYTctMDVlZC00MjczLWFkYTItOTRlNGNmMWY3N2MyIj4KICAgICAgICA8ZGM6Qm91bmRzIGhlaWdodD0iODAuMCIgd2lkdGg9IjEwMC4wIiB4PSIxODYuMCIgeT0iNzguMCIvPgogICAgICA8L2JwbW5kaTpCUE1OU2hhcGU+CiAgICAgIDxicG1uZGk6QlBNTkVkZ2UgYnBtbkVsZW1lbnQ9InNlcXVlbmNlRmxvd18yY2IyNTNjMS01ZjVmLTRmYmMtODQ4Yy0xN2I5MjgyMzc5ODIiIGlkPSJCUE1ORWRnZV85NDE3YzZjNS0xNDE1LTQ5ZjktYjAzOS01NzBiM2YyMTE4YmIiPgogICAgICAgIDxkaTp3YXlwb2ludCB4PSIxMzYuMCIgeT0iMTE4LjAiLz4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMTg2LjAiIHk9IjExOC4wIi8+CiAgICAgIDwvYnBtbmRpOkJQTU5FZGdlPgogICAgICA8YnBtbmRpOkJQTU5TaGFwZSBicG1uRWxlbWVudD0iZW5kRXZlbnRfM2IzNmNkZGQtYTJkYy00YTVjLTk4ODEtNzFkZTY4OTEzMjA3IiBpZD0iQlBNTlNoYXBlXzY5ZDUxMDMzLTExYmQtNGI0ZS05NGQ0LWIxM2NlYjBjYWQyYSI+CiAgICAgICAgPGRjOkJvdW5kcyBoZWlnaHQ9IjM2LjAiIHdpZHRoPSIzNi4wIiB4PSIzMzYuMCIgeT0iMTAwLjAiLz4KICAgICAgPC9icG1uZGk6QlBNTlNoYXBlPgogICAgICA8YnBtbmRpOkJQTU5FZGdlIGJwbW5FbGVtZW50PSJzZXF1ZW5jZUZsb3dfNGRkYjY0ZDAtYzU1ZC00YjY3LThmZDAtZTZkZWYwZGZiMDUzIiBpZD0iQlBNTkVkZ2VfMTNhMjU1OWItNzkzZC00ZGZlLWE4MDktZDg2YzVlMGQ3YmFkIj4KICAgICAgICA8ZGk6d2F5cG9pbnQgeD0iMjg2LjAiIHk9IjExOC4wIi8+CiAgICAgICAgPGRpOndheXBvaW50IHg9IjMzNi4wIiB5PSIxMTguMCIvPgogICAgICA8L2JwbW5kaTpCUE1ORWRnZT4KICAgIDwvYnBtbmRpOkJQTU5QbGFuZT4KICA8L2JwbW5kaTpCUE1ORGlhZ3JhbT4KPC9kZWZpbml0aW9ucz4K"}],"deployedWorkflows":[{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"resourceName":"process.xml"}]}
Position: 5 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE_CREATION, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","workflowKey":-1,"workflowInstanceKey":-1,"version":-1,"variables":"gA==","fetchVariables":[]}
Position: 6 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 7 Key: 2251799813685252
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE_CREATION, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"version":1,"variables":"gA==","fetchVariables":[]}
Position: 8 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 9 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 10 Key: 2251799813685253
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 11 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 12 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 13 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 14 Key: 2251799813685254
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_fb1df54a-7aaa-4e67-9d3b-9f697c308db5","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 15 Key: 2251799813685255
RecordMetadata{recordType=EVENT, intentValue=255, intent=SEQUENCE_FLOW_TAKEN, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"sequenceFlow_2cb253c1-5f5f-4fbc-848c-17b928237982","flowScopeKey":2251799813685251,"bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 16 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 17 Key: 2251799813685253
RecordMetadata{recordType=COMMAND, intentValue=255, intent=TRIGGER, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 18 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 19 Key: 2251799813685258
RecordMetadata{recordType=EVENT, intentValue=255, intent=EVENT_OCCURRED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685251,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 20 Key: 2251799813685253
RecordMetadata{recordType=EVENT, intentValue=255, intent=TRIGGERED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=TIMER, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"elementInstanceKey":2251799813685251,"workflowInstanceKey":2251799813685251,"dueDate":1611841141852,"targetElementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","repetitions":1,"workflowKey":2251799813685249}
Position: 21 Key: -1
RecordMetadata{recordType=COMMAND, intentValue=255, intent=CREATE, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=JOB, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"deadline":-1,"worker":"","retries":3,"type":"test","customHeaders":[packed value (length=1)],"variables":"gA==","errorMessage":"","errorCode":"","bpmnProcessId":"workflow","workflowDefinitionVersion":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","elementInstanceKey":2251799813685256}
Position: 22 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_TERMINATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 23 Key: 2251799813685260
RecordMetadata{recordType=EVENT, intentValue=255, intent=CREATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=JOB, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"deadline":-1,"worker":"","retries":3,"type":"test","customHeaders":[packed value (length=1)],"variables":"gA==","errorMessage":"","errorCode":"","bpmnProcessId":"workflow","workflowDefinitionVersion":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","elementInstanceKey":2251799813685256}
Position: 24 Key: 2251799813685256
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_TERMINATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"task","flowScopeKey":2251799813685251,"bpmnElementType":"SERVICE_TASK","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 25 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 26 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 27 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 28 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 29 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 30 Key: 2251799813685261
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"startEvent_4d0922ed-3cbc-444a-9f45-cdd97be132b1","flowScopeKey":2251799813685259,"bpmnElementType":"START_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 31 Key: 2251799813685262
RecordMetadata{recordType=EVENT, intentValue=255, intent=SEQUENCE_FLOW_TAKEN, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"sequenceFlow_878d5d38-89fa-4ba9-9a60-e8f2199032da","flowScopeKey":2251799813685259,"bpmnElementType":"SEQUENCE_FLOW","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 32 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 33 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_ACTIVATED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 34 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 35 Key: 2251799813685263
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"endEvent_9cac0b9c-c7f5-480f-bc92-258a1405da18","flowScopeKey":2251799813685259,"bpmnElementType":"END_EVENT","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 36 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 37 Key: 2251799813685259
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"event-subprocess","flowScopeKey":2251799813685251,"bpmnElementType":"SUB_PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 38 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETING, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
Position: 39 Key: 2251799813685251
RecordMetadata{recordType=EVENT, intentValue=255, intent=ELEMENT_COMPLETED, requestStreamId=-2147483648, requestId=-1, protocolVersion=3, valueType=WORKFLOW_INSTANCE, rejectionType=NULL_VAL, rejectionReason=, brokerVersion=0.27.0}
Value:
{"bpmnProcessId":"workflow","version":1,"workflowKey":2251799813685249,"workflowInstanceKey":2251799813685251,"elementId":"workflow","flowScopeKey":-1,"bpmnElementType":"PROCESS","parentWorkflowInstanceKey":-1,"parentElementInstanceKey":-1}
13:39:12.489 [] DEBUG io.zeebe.util.buffer - Close stream processor
13:39:12.490 [Broker-0-StreamProcessor-1] DEBUG io.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1.
13:39:12.588 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - Close appender for log stream stream-1
13:39:12.589 [stream-1-write-buffer] DEBUG io.zeebe.dispatcher - Dispatcher closed
13:39:12.590 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - On closing logstream stream-1 close 4 readers
13:39:12.590 [Broker-0-LogStream-1] INFO io.zeebe.logstreams - Close log storage with name stream-1
13:39:12.637 [] DEBUG io.zeebe.broker.test - Clean up test files on path /tmp/junit17217163829394622494
13:39:12.638 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers'
13:39:12.638 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-actors'
13:39:12.639 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers': closed successfully
13:39:12.640 [] DEBUG io.zeebe.util.actor - Closing actor thread ground '-zb-actors': closed successfully</pre>
</details>
| test | workflowexecutioncleanstatetest testworkflowwitheventsubprocess summary how often does the test fail seldom at least i haven t seen it before does it block your work no do we suspect that it is a real failure based on the assertion it might simply be a condition that took too long to be met but i didn t look into it failures outline known failure cases e g a failed assertion and its stacktrace obtained from jenkins assertion failure org awaitility core conditiontimeoutexception assertion condition defined as a lambda expression in io zeebe engine state workflowexecutioncleanstatetest expecting empty but was within seconds at org awaitility core conditionawaiter await conditionawaiter java at org awaitility core assertioncondition await assertioncondition java at org awaitility core assertioncondition await assertioncondition java at org awaitility core conditionfactory until conditionfactory java at org awaitility core conditionfactory untilasserted conditionfactory java at io zeebe engine state workflowexecutioncleanstatetest assertthatstateisempty workflowexecutioncleanstatetest java at io zeebe engine state workflowexecutioncleanstatetest testworkflowwitheventsubprocess workflowexecutioncleanstatetest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit rules testwatcher evaluate testwatcher java at org junit rules externalresource evaluate externalresource java at org junit rules testwatcher evaluate testwatcher java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runner junitcore run junitcore java at org junit runner junitcore run junitcore java at org junit vintage engine execution runnerexecutor execute runnerexecutor java at java base java util stream foreachops foreachop ofref accept foreachops java at java base java util stream referencepipeline accept referencepipeline java at java base java util iterator foreachremaining iterator java at java base java util spliterators iteratorspliterator foreachremaining spliterators java at java base java util stream abstractpipeline copyinto abstractpipeline java at java base java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java base java util stream foreachops foreachop evaluatesequential foreachops java at java base java util stream foreachops foreachop ofref evaluatesequential foreachops java at java base java util stream abstractpipeline evaluate abstractpipeline java at java base java util stream referencepipeline foreach referencepipeline java at org junit vintage engine vintagetestengine executeallchildren vintagetestengine java at org junit vintage engine vintagetestengine execute vintagetestengine java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator lambda execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator withinterceptedstreams engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org apache maven surefire junitplatform junitplatformprovider lambda execute junitplatformprovider java at java base java util iterator foreachremaining iterator java at org apache maven surefire junitplatform junitplatformprovider execute junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invokealltests junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invoke junitplatformprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java lang assertionerror expecting empty but was at io zeebe engine state workflowexecutioncleanstatetest lambda assertthatstateisempty workflowexecutioncleanstatetest java at org awaitility core assertioncondition lambda new assertioncondition java at org awaitility core conditionawaiter conditionpoller call conditionawaiter java at org awaitility core conditionawaiter conditionpoller call conditionawaiter java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java logs logs debug io zeebe logstreams configured log appender back pressure at partition as appendervegascfg initiallimit maxconcurrency alphalimit betalimit window limiting is disabled debug io zeebe logstreams recovering state of partition from snapshot info org camunda feel feelengine engine created info io zeebe logstreams recovered state of partition from snapshot at position info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda feel feelengine engine created info io zeebe test records test failed following records were exported info io zeebe test records valuetype deployment key position timestamp recordtype command intent create partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deployedworkflows resources resourcetype bpmn xml resourcename process xml resource info io zeebe test records valuetype deployment key position timestamp recordtype event intent created partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deployedworkflows resources resourcetype bpmn xml resourcename process xml resource info io zeebe test records valuetype deployment key position timestamp recordtype command intent distribute partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deployedworkflows resources resourcetype bpmn xml resourcename process xml resource info io zeebe test records valuetype deployment key position timestamp recordtype event intent distributed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deployedworkflows resources resourcetype bpmn xml resourcename process xml resource info io zeebe test records valuetype workflow instance creation key position timestamp recordtype command intent create partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value variables version workflowinstancekey bpmnprocessid workflow workflowkey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid workflow bpmnelementtype process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance creation key position timestamp recordtype event intent created partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value variables version workflowinstancekey bpmnprocessid workflow workflowkey info io zeebe test records valuetype timer key position timestamp recordtype command intent create partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value targetelementid startevent workflowinstancekey workflowkey repetitions duedate elementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid workflow bpmnelementtype process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype timer key position timestamp recordtype event intent created partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value targetelementid startevent workflowinstancekey workflowkey repetitions duedate elementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completing partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent sequence flow taken partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid sequenceflow bpmnelementtype sequence flow parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid task bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype timer key position timestamp recordtype command intent trigger partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value targetelementid startevent workflowinstancekey workflowkey repetitions duedate elementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid task bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent event occurred partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype timer key position timestamp recordtype event intent triggered partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value targetelementid startevent workflowinstancekey workflowkey repetitions duedate elementinstancekey info io zeebe test records valuetype job key position timestamp recordtype command intent create partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deadline variables errormessage type test workflowinstancekey bpmnprocessid workflow workflowkey elementid task elementinstancekey retries customheaders worker workflowdefinitionversion errorcode info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element terminating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid task bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype job key position timestamp recordtype event intent created partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value deadline variables errormessage type test workflowinstancekey bpmnprocessid workflow workflowkey elementid task elementinstancekey retries customheaders worker workflowdefinitionversion errorcode info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element terminated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid task bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid event subprocess bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid event subprocess bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completing partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid startevent bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent sequence flow taken partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid sequenceflow bpmnelementtype sequence flow parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activating partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid endevent bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element activated partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid endevent bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completing partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid endevent bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid endevent bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completing partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid event subprocess bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid event subprocess bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completing partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid workflow bpmnelementtype process parentworkflowinstancekey parentelementinstancekey info io zeebe test records valuetype workflow instance key position timestamp recordtype event intent element completed partitionid rejectiontype null val rejectionreason brokerversion sourcerecordposition value version flowscopekey workflowinstancekey bpmnprocessid workflow workflowkey elementid workflow bpmnelementtype process parentworkflowinstancekey parentelementinstancekey info io zeebe broker test test failed following records were exported info io zeebe broker test records on partition position key recordmetadata recordtype command intentvalue intent create requeststreamid requestid protocolversion valuetype deployment rejectiontype null val rejectionreason brokerversion value resources deployedworkflows position key recordmetadata recordtype event intentvalue intent created requeststreamid requestid protocolversion valuetype deployment rejectiontype null val rejectionreason brokerversion value resources deployedworkflows position key recordmetadata recordtype command intentvalue intent distribute requeststreamid requestid protocolversion valuetype deployment rejectiontype null val rejectionreason brokerversion value resources deployedworkflows position key recordmetadata recordtype event intentvalue intent distributed requeststreamid requestid protocolversion valuetype deployment rejectiontype null val rejectionreason brokerversion value resources deployedworkflows position key recordmetadata recordtype command intentvalue intent create requeststreamid requestid protocolversion valuetype workflow instance creation rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow workflowkey workflowinstancekey version variables ga fetchvariables position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid workflow flowscopekey bpmnelementtype process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent created requeststreamid requestid protocolversion valuetype workflow instance creation rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow workflowkey workflowinstancekey version variables ga fetchvariables position key recordmetadata recordtype command intentvalue intent create requeststreamid requestid protocolversion valuetype timer rejectiontype null val rejectionreason brokerversion value elementinstancekey workflowinstancekey duedate targetelementid startevent repetitions workflowkey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid workflow flowscopekey bpmnelementtype process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent created requeststreamid requestid protocolversion valuetype timer rejectiontype null val rejectionreason brokerversion value elementinstancekey workflowinstancekey duedate targetelementid startevent repetitions workflowkey position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completing requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completed requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent sequence flow taken requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid sequenceflow flowscopekey bpmnelementtype sequence flow parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid task flowscopekey bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype command intentvalue intent trigger requeststreamid requestid protocolversion valuetype timer rejectiontype null val rejectionreason brokerversion value elementinstancekey workflowinstancekey duedate targetelementid startevent repetitions workflowkey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid task flowscopekey bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent event occurred requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent triggered requeststreamid requestid protocolversion valuetype timer rejectiontype null val rejectionreason brokerversion value elementinstancekey workflowinstancekey duedate targetelementid startevent repetitions workflowkey position key recordmetadata recordtype command intentvalue intent create requeststreamid requestid protocolversion valuetype job rejectiontype null val rejectionreason brokerversion value deadline worker retries type test customheaders variables ga errormessage errorcode bpmnprocessid workflow workflowdefinitionversion workflowkey workflowinstancekey elementid task elementinstancekey position key recordmetadata recordtype event intentvalue intent element terminating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid task flowscopekey bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent created requeststreamid requestid protocolversion valuetype job rejectiontype null val rejectionreason brokerversion value deadline worker retries type test customheaders variables ga errormessage errorcode bpmnprocessid workflow workflowdefinitionversion workflowkey workflowinstancekey elementid task elementinstancekey position key recordmetadata recordtype event intentvalue intent element terminated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid task flowscopekey bpmnelementtype service task parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid event subprocess flowscopekey bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid event subprocess flowscopekey bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completing requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completed requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid startevent flowscopekey bpmnelementtype start event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent sequence flow taken requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid sequenceflow flowscopekey bpmnelementtype sequence flow parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activating requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid endevent flowscopekey bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element activated requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid endevent flowscopekey bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completing requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid endevent flowscopekey bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completed requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid endevent flowscopekey bpmnelementtype end event parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completing requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid event subprocess flowscopekey bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completed requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid event subprocess flowscopekey bpmnelementtype sub process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completing requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid workflow flowscopekey bpmnelementtype process parentworkflowinstancekey parentelementinstancekey position key recordmetadata recordtype event intentvalue intent element completed requeststreamid requestid protocolversion valuetype workflow instance rejectiontype null val rejectionreason brokerversion value bpmnprocessid workflow version workflowkey workflowinstancekey elementid workflow flowscopekey bpmnelementtype process parentworkflowinstancekey parentelementinstancekey debug io zeebe util buffer close stream processor debug io zeebe logstreams closed stream processor controller broker streamprocessor info io zeebe logstreams close appender for log stream stream debug io zeebe dispatcher dispatcher closed info io zeebe logstreams on closing logstream stream close readers info io zeebe logstreams close log storage with name stream debug io zeebe broker test clean up test files on path tmp debug io zeebe util actor closing actor thread ground zb fs workers debug io zeebe util actor closing actor thread ground zb actors debug io zeebe util actor closing actor thread ground zb fs workers closed successfully debug io zeebe util actor closing actor thread ground zb actors closed successfully | 1 |
279,251 | 24,208,469,837 | IssuesEvent | 2022-09-25 15:13:25 | moodle-an-hochschulen/moodle-theme_boost_union | https://api.github.com/repos/moodle-an-hochschulen/moodle-theme_boost_union | opened | Add missing Behat tests for Scroll-spy implementation | enhancement test | In the review of #80, it became obvious that there aren't any Behat tests for the scroll-spy feature yet.
There are two stub scenarios in theme_boost_union_feelsettings_navigation.feature, but these will have to be filled with real testing steps now. | 1.0 | Add missing Behat tests for Scroll-spy implementation - In the review of #80, it became obvious that there aren't any Behat tests for the scroll-spy feature yet.
There are two stub scenarios in theme_boost_union_feelsettings_navigation.feature, but these will have to be filled with real testing steps now. | test | add missing behat tests for scroll spy implementation in the review of it became obvious that there aren t any behat tests for the scroll spy feature yet there are two stub scenarios in theme boost union feelsettings navigation feature but these will have to be filled with real testing steps now | 1 |
311,028 | 26,762,345,676 | IssuesEvent | 2023-01-31 08:06:47 | saleor/saleor-dashboard | https://api.github.com/repos/saleor/saleor-dashboard | closed | Cypress test fail: should be able to create variant visible for the customers in all channels. TC: SALEOR_2901 | tests | **Known bug for versions:**
v37: false
**Additional Info:**
Spec: As an admin I should be able to create variant
**Created ticket**
For: QA
Link: https://github.com/saleor/saleor-dashboard/issues/2781 | 1.0 | Cypress test fail: should be able to create variant visible for the customers in all channels. TC: SALEOR_2901 - **Known bug for versions:**
v37: false
**Additional Info:**
Spec: As an admin I should be able to create variant
**Created ticket**
For: QA
Link: https://github.com/saleor/saleor-dashboard/issues/2781 | test | cypress test fail should be able to create variant visible for the customers in all channels tc saleor known bug for versions false additional info spec as an admin i should be able to create variant created ticket for qa link | 1 |
65,976 | 12,701,779,486 | IssuesEvent | 2020-06-22 18:50:21 | AlaskaAirlines/auro_docs | https://api.github.com/repos/AlaskaAirlines/auro_docs | opened | Footer | Artwork: ADS Compliant Code: Backlogged Docs: Backlogged UI Kit: Backlogged | # Component Audit
The scope of this issue is to track the full progress of the Footer component.
<!-- template content, do not remove -->
## Process
This issue is to remain open until all areas of the specification and all assets have been delivered to stakeholders and user groups.
As new issues are created that are children of this Audit item, please be sure to reference this Audit item in that issue.
As assets are delivered, it is required that the labels associated with this Audit issue are updated to reflect status. Labels are as follows:
- Artwork
- Code
- Docs
- Ui Kit
This Audit issue will remain open until all assets have been delivered be it V0 or ADS compliment.
| 1.0 | Footer - # Component Audit
The scope of this issue is to track the full progress of the Footer component.
<!-- template content, do not remove -->
## Process
This issue is to remain open until all areas of the specification and all assets have been delivered to stakeholders and user groups.
As new issues are created that are children of this Audit item, please be sure to reference this Audit item in that issue.
As assets are delivered, it is required that the labels associated with this Audit issue are updated to reflect status. Labels are as follows:
- Artwork
- Code
- Docs
- Ui Kit
This Audit issue will remain open until all assets have been delivered be it V0 or ADS compliment.
| non_test | footer component audit the scope of this issue is to track the full progress of the footer component process this issue is to remain open until all areas of the specification and all assets have been delivered to stakeholders and user groups as new issues are created that are children of this audit item please be sure to reference this audit item in that issue as assets are delivered it is required that the labels associated with this audit issue are updated to reflect status labels are as follows artwork code docs ui kit this audit issue will remain open until all assets have been delivered be it or ads compliment | 0 |
296,486 | 25,553,268,519 | IssuesEvent | 2022-11-30 02:52:56 | apache/apisix-ingress-controller | https://api.github.com/repos/apache/apisix-ingress-controller | closed | test: ensure CI robustness | area/test stale | ## backend
ref: #1138
Due to the instability of E2E on CI, the development efficiency and time cost of contributors are greatly affected. In this issue, the cause of e2e instability is mainly recorded and determined.
**Main problems**
* Use incorrect methods to ensure that resources are correctly written to `apisix`.
* Sleep does not have stability factors, so we need to implement a method to "ensure the correct synchronization of resources to apisix" to replace it.
* There are bug in `ingress-controller` itself. (undetermined)
**error list**
- [x] `endpoints.go` #1149
- [x] `enable plugin and then delete it` #1149
- [x] `resourcesync.go` #1149
- [ ] ... | 1.0 | test: ensure CI robustness - ## backend
ref: #1138
Due to the instability of E2E on CI, the development efficiency and time cost of contributors are greatly affected. In this issue, the cause of e2e instability is mainly recorded and determined.
**Main problems**
* Use incorrect methods to ensure that resources are correctly written to `apisix`.
* Sleep does not have stability factors, so we need to implement a method to "ensure the correct synchronization of resources to apisix" to replace it.
* There are bug in `ingress-controller` itself. (undetermined)
**error list**
- [x] `endpoints.go` #1149
- [x] `enable plugin and then delete it` #1149
- [x] `resourcesync.go` #1149
- [ ] ... | test | test ensure ci robustness backend ref due to the instability of on ci the development efficiency and time cost of contributors are greatly affected in this issue the cause of instability is mainly recorded and determined main problems use incorrect methods to ensure that resources are correctly written to apisix sleep does not have stability factors so we need to implement a method to ensure the correct synchronization of resources to apisix to replace it there are bug in ingress controller itself undetermined error list endpoints go enable plugin and then delete it resourcesync go | 1 |
305,994 | 23,140,389,298 | IssuesEvent | 2022-07-28 17:53:37 | ESCOMP/CTSM | https://api.github.com/repos/ESCOMP/CTSM | closed | Add an IVT column to Table 2.2.1 as is done in Table 2.26.1 in Tech Note | type: documentation | User's are confused about the indices for pfts used in the surface dataset. We can clarify by adding an IVT column to Table 2.2.1 (as is done for crop pfts in Table 2.26.1). And include bare soil which is IVT=0. | 1.0 | Add an IVT column to Table 2.2.1 as is done in Table 2.26.1 in Tech Note - User's are confused about the indices for pfts used in the surface dataset. We can clarify by adding an IVT column to Table 2.2.1 (as is done for crop pfts in Table 2.26.1). And include bare soil which is IVT=0. | non_test | add an ivt column to table as is done in table in tech note user s are confused about the indices for pfts used in the surface dataset we can clarify by adding an ivt column to table as is done for crop pfts in table and include bare soil which is ivt | 0 |
16,914 | 3,574,938,556 | IssuesEvent | 2016-01-27 14:11:03 | Microsoft/vscode | https://api.github.com/repos/Microsoft/vscode | closed | Horizontal panel - general | testplan-item workbench | - [x] win @dbaeumer
- [x] mac @egamma
- [x] linux @jrieken
We introduced a concept of the horizontal panel to the workbench. For specific output and repl functionality there is #2176, so no need to test that. Verify:
* You can **toggle** the panel, either from the view menu or using the shortcuts
* You can nicely **resize** the horizontal panel using the sash which causes smooth layouting
* When opened horizontal panel always **restores** it's size and opens the last active panel (repl, git output, task output)
* Horizontal panel has a **maximum** size, and a **minimum** size below which it is minimized like the sidebar
* Be creative and try to **break** the workbench layout by minimizing / maximizing components, moving sidebar to the right, toggling visibility... | 1.0 | Horizontal panel - general - - [x] win @dbaeumer
- [x] mac @egamma
- [x] linux @jrieken
We introduced a concept of the horizontal panel to the workbench. For specific output and repl functionality there is #2176, so no need to test that. Verify:
* You can **toggle** the panel, either from the view menu or using the shortcuts
* You can nicely **resize** the horizontal panel using the sash which causes smooth layouting
* When opened horizontal panel always **restores** it's size and opens the last active panel (repl, git output, task output)
* Horizontal panel has a **maximum** size, and a **minimum** size below which it is minimized like the sidebar
* Be creative and try to **break** the workbench layout by minimizing / maximizing components, moving sidebar to the right, toggling visibility... | test | horizontal panel general win dbaeumer mac egamma linux jrieken we introduced a concept of the horizontal panel to the workbench for specific output and repl functionality there is so no need to test that verify you can toggle the panel either from the view menu or using the shortcuts you can nicely resize the horizontal panel using the sash which causes smooth layouting when opened horizontal panel always restores it s size and opens the last active panel repl git output task output horizontal panel has a maximum size and a minimum size below which it is minimized like the sidebar be creative and try to break the workbench layout by minimizing maximizing components moving sidebar to the right toggling visibility | 1 |
194,350 | 15,420,878,949 | IssuesEvent | 2021-03-05 12:14:37 | knurling-rs/defmt | https://api.github.com/repos/knurling-rs/defmt | closed | defmt.ferrous-systems.com is outdated (?) | type: documentation | at least the [timestamp section](https://defmt.ferrous-systems.com/timestamp.html) still documents the `#[defmt::timestamp]` attribute. | 1.0 | defmt.ferrous-systems.com is outdated (?) - at least the [timestamp section](https://defmt.ferrous-systems.com/timestamp.html) still documents the `#[defmt::timestamp]` attribute. | non_test | defmt ferrous systems com is outdated at least the still documents the attribute | 0 |
361,183 | 25,329,223,019 | IssuesEvent | 2022-11-18 11:52:21 | discopop-project/discopop | https://api.github.com/repos/discopop-project/discopop | closed | Split explanations for do-all and reduction patterns | documentation | Currently, it is mentioned in the wiki that reduction patterns are reported via do-all patterns, which is incorrect. | 1.0 | Split explanations for do-all and reduction patterns - Currently, it is mentioned in the wiki that reduction patterns are reported via do-all patterns, which is incorrect. | non_test | split explanations for do all and reduction patterns currently it is mentioned in the wiki that reduction patterns are reported via do all patterns which is incorrect | 0 |
132,562 | 18,740,958,007 | IssuesEvent | 2021-11-04 13:33:37 | hashicorp/terraform | https://api.github.com/repos/hashicorp/terraform | closed | 0.14+ throws "count" value depends on resource attributes that cannot be determined until apply | working as designed explained | <!--
Hi there,
Thank you for opening an issue. Please note that we try to keep the Terraform issue tracker reserved for bug reports and feature requests. For general usage questions, please see: https://www.terraform.io/community.html.
If your issue relates to Terraform Cloud/Enterprise, please contact tf-cloud@hashicorp.support.
If your issue relates to a specific Terraform provider, please open it in the provider's own repository. The index of providers is at https://registry.terraform.io/browse/providers.
To fix problems, we need clear reproduction cases - we need to be able to see it happen locally. A reproduction case is ideally something a Terraform Core engineer can git-clone or copy-paste and run immediately, without inventing any details or context.
* A short example can be directly copy-pasteable; longer examples should be in separate git repositories, especially if multiple files are needed
* Please include all needed context. For example, if you figured out that an expression can cause a crash, put the expression in a variable definition or a resource
* Set defaults on (or omit) any variables. The person reproducing it should not need to invent variable settings
* If multiple steps are required, such as running terraform twice, consider scripting it in a simple shell script. Providing a script can be easier than explaining what changes to make to the config between runs.
* Omit any unneeded complexity: remove variables, conditional statements, functions, modules, providers, and resources that are not needed to trigger the bug
* When possible, use the [null resource](https://www.terraform.io/docs/providers/null/resource.html) provider rather than a real provider in order to minimize external dependencies. We know this isn't always feasible. The Terraform Core team doesn't have deep domain knowledge in every provider, or access to every cloud platform for reproduction cases.
-->
### Terraform Version
```
v0.14.11 (and v1.0.6)
```
### Terraform Configuration Files
A bit of a contrived example to simplify repro but real code where error is [here](https://github.com/terraform-google-modules/terraform-google-project-factory/blob/e1c6a534ff8f69343edb839dc2496d108483ebd3/modules/shared_vpc_access/main.tf#L48)
```terraform
locals {
kv = {
"mykey1" : "foo",
"mykey2" : "bar",
}
tomatch = ["mykey1", "mykey3"]
matches = setintersection(keys(local.kv), local.tomatch)
transformed = length(local.random_res) != 0 ? tolist(setproduct(local.matches, local.random_res)) : []
random_res = [for r in random_pet.rands : r.id]
}
resource "random_pet" "rands" {
for_each = toset(["foo", "bar"])
}
resource "random_pet" "other" {
count = length(local.transformed)
}
```
### Debug Output
<!--
Full debug output can be obtained by running Terraform with the environment variable `TF_LOG=trace`. Please create a GitHub Gist containing the debug output. Please do _not_ paste the debug output in the issue, since debug output is long.
Debug output may contain sensitive information. Please review it before posting publicly, and if you are concerned feel free to encrypt the files using the HashiCorp security public key.
-->
### Expected Behavior
apply works (as with v0.13.7)
### Actual Behavior
The "count" value depends on resource attributes that cannot be determined
until apply, so Terraform cannot predict how many instances will be created.
To work around this, use the -target argument to first apply only the
resources that the count depends on.
### Steps to Reproduce
1. `terraform init`
2. `terraform apply`
### Additional Context
<!--
Are there anything atypical about your situation that we should know? For example: is Terraform running in a wrapper script or in a CI system? Are you passing any unusual command line options or environment variables to opt-in to non-default behavior?
-->
### References
<!--
Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here? For example:
- #6017
-->
| 1.0 | 0.14+ throws "count" value depends on resource attributes that cannot be determined until apply - <!--
Hi there,
Thank you for opening an issue. Please note that we try to keep the Terraform issue tracker reserved for bug reports and feature requests. For general usage questions, please see: https://www.terraform.io/community.html.
If your issue relates to Terraform Cloud/Enterprise, please contact tf-cloud@hashicorp.support.
If your issue relates to a specific Terraform provider, please open it in the provider's own repository. The index of providers is at https://registry.terraform.io/browse/providers.
To fix problems, we need clear reproduction cases - we need to be able to see it happen locally. A reproduction case is ideally something a Terraform Core engineer can git-clone or copy-paste and run immediately, without inventing any details or context.
* A short example can be directly copy-pasteable; longer examples should be in separate git repositories, especially if multiple files are needed
* Please include all needed context. For example, if you figured out that an expression can cause a crash, put the expression in a variable definition or a resource
* Set defaults on (or omit) any variables. The person reproducing it should not need to invent variable settings
* If multiple steps are required, such as running terraform twice, consider scripting it in a simple shell script. Providing a script can be easier than explaining what changes to make to the config between runs.
* Omit any unneeded complexity: remove variables, conditional statements, functions, modules, providers, and resources that are not needed to trigger the bug
* When possible, use the [null resource](https://www.terraform.io/docs/providers/null/resource.html) provider rather than a real provider in order to minimize external dependencies. We know this isn't always feasible. The Terraform Core team doesn't have deep domain knowledge in every provider, or access to every cloud platform for reproduction cases.
-->
### Terraform Version
```
v0.14.11 (and v1.0.6)
```
### Terraform Configuration Files
A bit of a contrived example to simplify repro but real code where error is [here](https://github.com/terraform-google-modules/terraform-google-project-factory/blob/e1c6a534ff8f69343edb839dc2496d108483ebd3/modules/shared_vpc_access/main.tf#L48)
```terraform
locals {
kv = {
"mykey1" : "foo",
"mykey2" : "bar",
}
tomatch = ["mykey1", "mykey3"]
matches = setintersection(keys(local.kv), local.tomatch)
transformed = length(local.random_res) != 0 ? tolist(setproduct(local.matches, local.random_res)) : []
random_res = [for r in random_pet.rands : r.id]
}
resource "random_pet" "rands" {
for_each = toset(["foo", "bar"])
}
resource "random_pet" "other" {
count = length(local.transformed)
}
```
### Debug Output
<!--
Full debug output can be obtained by running Terraform with the environment variable `TF_LOG=trace`. Please create a GitHub Gist containing the debug output. Please do _not_ paste the debug output in the issue, since debug output is long.
Debug output may contain sensitive information. Please review it before posting publicly, and if you are concerned feel free to encrypt the files using the HashiCorp security public key.
-->
### Expected Behavior
apply works (as with v0.13.7)
### Actual Behavior
The "count" value depends on resource attributes that cannot be determined
until apply, so Terraform cannot predict how many instances will be created.
To work around this, use the -target argument to first apply only the
resources that the count depends on.
### Steps to Reproduce
1. `terraform init`
2. `terraform apply`
### Additional Context
<!--
Are there anything atypical about your situation that we should know? For example: is Terraform running in a wrapper script or in a CI system? Are you passing any unusual command line options or environment variables to opt-in to non-default behavior?
-->
### References
<!--
Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here? For example:
- #6017
-->
| non_test | throws count value depends on resource attributes that cannot be determined until apply hi there thank you for opening an issue please note that we try to keep the terraform issue tracker reserved for bug reports and feature requests for general usage questions please see if your issue relates to terraform cloud enterprise please contact tf cloud hashicorp support if your issue relates to a specific terraform provider please open it in the provider s own repository the index of providers is at to fix problems we need clear reproduction cases we need to be able to see it happen locally a reproduction case is ideally something a terraform core engineer can git clone or copy paste and run immediately without inventing any details or context a short example can be directly copy pasteable longer examples should be in separate git repositories especially if multiple files are needed please include all needed context for example if you figured out that an expression can cause a crash put the expression in a variable definition or a resource set defaults on or omit any variables the person reproducing it should not need to invent variable settings if multiple steps are required such as running terraform twice consider scripting it in a simple shell script providing a script can be easier than explaining what changes to make to the config between runs omit any unneeded complexity remove variables conditional statements functions modules providers and resources that are not needed to trigger the bug when possible use the provider rather than a real provider in order to minimize external dependencies we know this isn t always feasible the terraform core team doesn t have deep domain knowledge in every provider or access to every cloud platform for reproduction cases terraform version and terraform configuration files a bit of a contrived example to simplify repro but real code where error is terraform locals kv foo bar tomatch matches setintersection keys local kv local tomatch transformed length local random res tolist setproduct local matches local random res random res resource random pet rands for each toset resource random pet other count length local transformed debug output full debug output can be obtained by running terraform with the environment variable tf log trace please create a github gist containing the debug output please do not paste the debug output in the issue since debug output is long debug output may contain sensitive information please review it before posting publicly and if you are concerned feel free to encrypt the files using the hashicorp security public key expected behavior apply works as with actual behavior the count value depends on resource attributes that cannot be determined until apply so terraform cannot predict how many instances will be created to work around this use the target argument to first apply only the resources that the count depends on steps to reproduce terraform init terraform apply additional context are there anything atypical about your situation that we should know for example is terraform running in a wrapper script or in a ci system are you passing any unusual command line options or environment variables to opt in to non default behavior references are there any other github issues open or closed or pull requests that should be linked here for example | 0 |
147,421 | 11,786,995,423 | IssuesEvent | 2020-03-17 13:19:07 | ClaudiaLapalme/pikaroute | https://api.github.com/repos/ClaudiaLapalme/pikaroute | closed | ST-6 Routes Generation | System Test | **User Story**
#58
**Steps**
1. Open the application
2. Open the directions page
3. Enter Parc Jean-Drapeau as From
4. Enter Berri-UQAM as to
5. Press Get Directions
**Expected Output**
Time cannot be checked as it depends on the time of the day the test is ran and the traffic
1. Walking should have one route
2. Public transport should have two routes
3. Cycling should have one route
4. Driving should have one route
**Result**
PASS
**Comments**
Due to limitations involving the Selenium IDE, it is currently impossible to do negative testing. ie we cannot check that a third route does not exist for public transport. | 1.0 | ST-6 Routes Generation - **User Story**
#58
**Steps**
1. Open the application
2. Open the directions page
3. Enter Parc Jean-Drapeau as From
4. Enter Berri-UQAM as to
5. Press Get Directions
**Expected Output**
Time cannot be checked as it depends on the time of the day the test is ran and the traffic
1. Walking should have one route
2. Public transport should have two routes
3. Cycling should have one route
4. Driving should have one route
**Result**
PASS
**Comments**
Due to limitations involving the Selenium IDE, it is currently impossible to do negative testing. ie we cannot check that a third route does not exist for public transport. | test | st routes generation user story steps open the application open the directions page enter parc jean drapeau as from enter berri uqam as to press get directions expected output time cannot be checked as it depends on the time of the day the test is ran and the traffic walking should have one route public transport should have two routes cycling should have one route driving should have one route result pass comments due to limitations involving the selenium ide it is currently impossible to do negative testing ie we cannot check that a third route does not exist for public transport | 1 |
300,413 | 25,966,880,300 | IssuesEvent | 2022-12-19 07:55:56 | cuny-academic-commons/commons-in-a-box | https://api.github.com/repos/cuny-academic-commons/commons-in-a-box | closed | CBOX OpenLab dev site update error | OpenLab Needs testing | Hi Boone,
I went to the dashboard on openlabdev.commonsinabox.org to check something for the ticket I just added, and it started automatically updating plugins. All updated successfully except OpenLab Attributions. I got the following error (just part of the text):
> An error occurred while updating OpenLab Attributions: The update cannot be installed because some files could not be copied. This is usually due to inconsistent file permissions. .git, .git/branches, .git/hooks, .git/hooks/applypatch-msg.sample, .git/hooks/commit-msg.sample, .git/hooks/fsmonitor-watchman.sample, .git/hooks/post-update.sample, .git/hooks/pre-applypatch.sample, .git/hooks/pre-commit.sample, .git/hooks/pre-merge-commit.sample, .git/hooks/pre-push.sample, .git/hooks/pre-rebase.sample, .git/hooks/pre-receive.sample, .git/hooks/prepare-commit-msg.sample, .git/hooks/push-to-checkout.sample, .git/hooks/update.sample, .git/info, .git/info/exclude, .git/logs, .git/logs/refs, .git/logs/refs/heads, .git/logs/refs/heads/2.0.x, .git/logs/refs/heads/build, .git/logs/refs/heads/main, .git/logs/refs/remotes, .git/logs/refs/remotes/origin, .git/logs/refs/remotes/origin/2.0.x, .git/logs/refs/remotes/origin/HEAD, .git/logs/refs/remotes/origin/build, .git/logs/refs/remotes/origin/main, .git/logs/HEAD, .git/objects, .git/objects/00, .git/objects/00/076cbd4bdf997556290319639cd3ce1c351b18, .git/objects/00/07e039b524a7b7afd95b0d4c72483b11bee999, .git/objects/00/885ff7c66f32a3fe209732f1c8c93d062236a7...
Now when I try to get back to the Dashboard, it keeps attempting to update that plugin so I just logged out. I'm not testing anything, so no rush, but just fyi .
Thanks!
| 1.0 | CBOX OpenLab dev site update error - Hi Boone,
I went to the dashboard on openlabdev.commonsinabox.org to check something for the ticket I just added, and it started automatically updating plugins. All updated successfully except OpenLab Attributions. I got the following error (just part of the text):
> An error occurred while updating OpenLab Attributions: The update cannot be installed because some files could not be copied. This is usually due to inconsistent file permissions. .git, .git/branches, .git/hooks, .git/hooks/applypatch-msg.sample, .git/hooks/commit-msg.sample, .git/hooks/fsmonitor-watchman.sample, .git/hooks/post-update.sample, .git/hooks/pre-applypatch.sample, .git/hooks/pre-commit.sample, .git/hooks/pre-merge-commit.sample, .git/hooks/pre-push.sample, .git/hooks/pre-rebase.sample, .git/hooks/pre-receive.sample, .git/hooks/prepare-commit-msg.sample, .git/hooks/push-to-checkout.sample, .git/hooks/update.sample, .git/info, .git/info/exclude, .git/logs, .git/logs/refs, .git/logs/refs/heads, .git/logs/refs/heads/2.0.x, .git/logs/refs/heads/build, .git/logs/refs/heads/main, .git/logs/refs/remotes, .git/logs/refs/remotes/origin, .git/logs/refs/remotes/origin/2.0.x, .git/logs/refs/remotes/origin/HEAD, .git/logs/refs/remotes/origin/build, .git/logs/refs/remotes/origin/main, .git/logs/HEAD, .git/objects, .git/objects/00, .git/objects/00/076cbd4bdf997556290319639cd3ce1c351b18, .git/objects/00/07e039b524a7b7afd95b0d4c72483b11bee999, .git/objects/00/885ff7c66f32a3fe209732f1c8c93d062236a7...
Now when I try to get back to the Dashboard, it keeps attempting to update that plugin so I just logged out. I'm not testing anything, so no rush, but just fyi .
Thanks!
| test | cbox openlab dev site update error hi boone i went to the dashboard on openlabdev commonsinabox org to check something for the ticket i just added and it started automatically updating plugins all updated successfully except openlab attributions i got the following error just part of the text an error occurred while updating openlab attributions the update cannot be installed because some files could not be copied this is usually due to inconsistent file permissions git git branches git hooks git hooks applypatch msg sample git hooks commit msg sample git hooks fsmonitor watchman sample git hooks post update sample git hooks pre applypatch sample git hooks pre commit sample git hooks pre merge commit sample git hooks pre push sample git hooks pre rebase sample git hooks pre receive sample git hooks prepare commit msg sample git hooks push to checkout sample git hooks update sample git info git info exclude git logs git logs refs git logs refs heads git logs refs heads x git logs refs heads build git logs refs heads main git logs refs remotes git logs refs remotes origin git logs refs remotes origin x git logs refs remotes origin head git logs refs remotes origin build git logs refs remotes origin main git logs head git objects git objects git objects git objects git objects now when i try to get back to the dashboard it keeps attempting to update that plugin so i just logged out i m not testing anything so no rush but just fyi thanks | 1 |
140,920 | 11,383,415,554 | IssuesEvent | 2020-01-29 05:56:09 | intellij-rust/intellij-rust | https://api.github.com/repos/intellij-rust/intellij-rust | closed | Detect #[tokio::test] attribute as testable function | feature subsystem:tests | ## Feature request
Tokio 0.2.x provides the [tokio::test](https://docs.rs/tokio/0.2.4/tokio/attr.test.html) attribute for defining async tests. The plugin does not detect these as a testable functions, i.e. no green "run" arrow is shown in the IDE (tested in CLion).
The `#[tokio::test]` attribute is implemented as a procedural-macro which ultimately outputs a standard `#[test]` attribute (see [here](https://docs.rs/tokio-macros/0.2.0/src/tokio_macros/lib.rs.html#201)). Given the plugin does not currently expand procedural macros it makes sense that it would not detect these.
Assuming support for procedural macro expansion is not imminent, it may be helpful to expand the detection logic for testable functions to cover attributes matching the pattern `#[.*::test]`.
| 1.0 | Detect #[tokio::test] attribute as testable function - ## Feature request
Tokio 0.2.x provides the [tokio::test](https://docs.rs/tokio/0.2.4/tokio/attr.test.html) attribute for defining async tests. The plugin does not detect these as a testable functions, i.e. no green "run" arrow is shown in the IDE (tested in CLion).
The `#[tokio::test]` attribute is implemented as a procedural-macro which ultimately outputs a standard `#[test]` attribute (see [here](https://docs.rs/tokio-macros/0.2.0/src/tokio_macros/lib.rs.html#201)). Given the plugin does not currently expand procedural macros it makes sense that it would not detect these.
Assuming support for procedural macro expansion is not imminent, it may be helpful to expand the detection logic for testable functions to cover attributes matching the pattern `#[.*::test]`.
| test | detect attribute as testable function feature request tokio x provides the attribute for defining async tests the plugin does not detect these as a testable functions i e no green run arrow is shown in the ide tested in clion the attribute is implemented as a procedural macro which ultimately outputs a standard attribute see given the plugin does not currently expand procedural macros it makes sense that it would not detect these assuming support for procedural macro expansion is not imminent it may be helpful to expand the detection logic for testable functions to cover attributes matching the pattern | 1 |
508,484 | 14,701,265,845 | IssuesEvent | 2021-01-04 11:34:32 | ramsayinnovations/ramsay-site | https://api.github.com/repos/ramsayinnovations/ramsay-site | closed | Create GoogleTagManager Component | Priority Review | Please create the following component in the `components` folder: https://github.com/vercel/next.js/blob/canary/examples/with-google-tag-manager/components/GoogleTagManager.js
This should be done in the `develop--clone` branch
| 1.0 | Create GoogleTagManager Component - Please create the following component in the `components` folder: https://github.com/vercel/next.js/blob/canary/examples/with-google-tag-manager/components/GoogleTagManager.js
This should be done in the `develop--clone` branch
| non_test | create googletagmanager component please create the following component in the components folder this should be done in the develop clone branch | 0 |
229,144 | 7,571,829,317 | IssuesEvent | 2018-04-23 13:25:38 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | After failed upload note should be blue | Enhancement Medium Priority | To reproduce:
- close note
- enable WiFi to allow upload
- be outside WiFi range to ensure upload failure
- upload note
- upload fails, note should be blue as it is waiting for upload
- note is green | 1.0 | After failed upload note should be blue - To reproduce:
- close note
- enable WiFi to allow upload
- be outside WiFi range to ensure upload failure
- upload note
- upload fails, note should be blue as it is waiting for upload
- note is green | non_test | after failed upload note should be blue to reproduce close note enable wifi to allow upload be outside wifi range to ensure upload failure upload note upload fails note should be blue as it is waiting for upload note is green | 0 |
244,528 | 20,675,069,082 | IssuesEvent | 2022-03-10 08:26:37 | ipfs/go-bitswap | https://api.github.com/repos/ipfs/go-bitswap | closed | TestSessionBetweenPeers is flaky | topic/test failure kind/test need/triage | Failed
=== RUN TestSessionBetweenPeers
TestSessionBetweenPeers: bitswap_with_sessions_test.go:127: uninvolved nodes should only receive two messages 4
--- FAIL: TestSessionBetweenPeers (0.89s) | 2.0 | TestSessionBetweenPeers is flaky - Failed
=== RUN TestSessionBetweenPeers
TestSessionBetweenPeers: bitswap_with_sessions_test.go:127: uninvolved nodes should only receive two messages 4
--- FAIL: TestSessionBetweenPeers (0.89s) | test | testsessionbetweenpeers is flaky failed run testsessionbetweenpeers testsessionbetweenpeers bitswap with sessions test go uninvolved nodes should only receive two messages fail testsessionbetweenpeers | 1 |
251,788 | 21,523,097,168 | IssuesEvent | 2022-04-28 15:46:48 | damccorm/test-migration-target | https://api.github.com/repos/damccorm/test-migration-target | opened | Java Examples are failing for different runners | test P2 testing examples-java | With the recent addition of suites to continuously test Java examples in different runners, some existing integration tests fail for different runners
* {_}org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInBatchDynamicSharding{_}, fails in FlinkRunner with AssertionError, for DirectRunner and SparkRunner the test never finishes.
* _org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInStreamingStaticSharding_ fails in SparkRunner with error _Failed to read from sharded output_
* _org.apache.beam.examples.cookbook.BigQueryTornadoesIT.testE2eBigQueryTornadoesWithStorageApiUsingQuery_ fails in FlinkRunner with _Error deleting table, Not found: Dataset_
Imported from Jira [BEAM-14019](https://issues.apache.org/jira/browse/BEAM-14019)
Reported by: benglez. | 2.0 | Java Examples are failing for different runners - With the recent addition of suites to continuously test Java examples in different runners, some existing integration tests fail for different runners
* {_}org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInBatchDynamicSharding{_}, fails in FlinkRunner with AssertionError, for DirectRunner and SparkRunner the test never finishes.
* _org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInStreamingStaticSharding_ fails in SparkRunner with error _Failed to read from sharded output_
* _org.apache.beam.examples.cookbook.BigQueryTornadoesIT.testE2eBigQueryTornadoesWithStorageApiUsingQuery_ fails in FlinkRunner with _Error deleting table, Not found: Dataset_
Imported from Jira [BEAM-14019](https://issues.apache.org/jira/browse/BEAM-14019)
Reported by: benglez. | test | java examples are failing for different runners with the recent addition of suites to continuously test java examples in different runners some existing integration tests fail for different runners org apache beam examples windowedwordcountit testwindowedwordcountinbatchdynamicsharding fails in flinkrunner with assertionerror for directrunner and sparkrunner the test never finishes org apache beam examples windowedwordcountit testwindowedwordcountinstreamingstaticsharding fails in sparkrunner with error failed to read from sharded output org apache beam examples cookbook bigquerytornadoesit fails in flinkrunner with error deleting table not found dataset imported from jira reported by benglez | 1 |
228,278 | 18,168,169,874 | IssuesEvent | 2021-09-27 16:44:18 | aces/Loris | https://api.github.com/repos/aces/Loris | opened | [genomic_browser] The CSV download is showing more information that the one on the UI (some with potentinal privacy issues) | Bug 24.0.0-testing | **Describe the bug**
The Genomic Browser CSV download is showing more information that the one on the UI (some with potential privacy issues).
**To Reproduce**
Steps to reproduce the behavior (attach screenshots if applicable):
1. Go to MainMenu->Genomics->Genomic Browser
2. Go to any of the tabs contaning information (row entries). (Profiles for example)
3. Got the the Download table as CSV and download the table.
4. There is more information show that the one filtered.
**IMPORTANT: Some of this information poses potential privacy issues (Date of birth for example)**

**What did you expect to happen?**
According to the test Plan only the info shown in the UI should be downloaded.
**Browser Environment (please complete the following information):**
- MSWindows 10 as guest. Site hosted on test Loris.
- Browser Firefox 92.0.1 (64-bit) and Chrome 94.0.4606.61 (Build oficial) (64 bits)
**Server Environment (if known):**
_Note: We only support the most recent release of LORIS._
- LORIS Version: [e.g. 24.0.0]
- Linux distribution and Version: Site hosted on test Loris
| 1.0 | [genomic_browser] The CSV download is showing more information that the one on the UI (some with potentinal privacy issues) - **Describe the bug**
The Genomic Browser CSV download is showing more information that the one on the UI (some with potential privacy issues).
**To Reproduce**
Steps to reproduce the behavior (attach screenshots if applicable):
1. Go to MainMenu->Genomics->Genomic Browser
2. Go to any of the tabs contaning information (row entries). (Profiles for example)
3. Got the the Download table as CSV and download the table.
4. There is more information show that the one filtered.
**IMPORTANT: Some of this information poses potential privacy issues (Date of birth for example)**

**What did you expect to happen?**
According to the test Plan only the info shown in the UI should be downloaded.
**Browser Environment (please complete the following information):**
- MSWindows 10 as guest. Site hosted on test Loris.
- Browser Firefox 92.0.1 (64-bit) and Chrome 94.0.4606.61 (Build oficial) (64 bits)
**Server Environment (if known):**
_Note: We only support the most recent release of LORIS._
- LORIS Version: [e.g. 24.0.0]
- Linux distribution and Version: Site hosted on test Loris
| test | the csv download is showing more information that the one on the ui some with potentinal privacy issues describe the bug the genomic browser csv download is showing more information that the one on the ui some with potential privacy issues to reproduce steps to reproduce the behavior attach screenshots if applicable go to mainmenu genomics genomic browser go to any of the tabs contaning information row entries profiles for example got the the download table as csv and download the table there is more information show that the one filtered important some of this information poses potential privacy issues date of birth for example what did you expect to happen according to the test plan only the info shown in the ui should be downloaded browser environment please complete the following information mswindows as guest site hosted on test loris browser firefox bit and chrome build oficial bits server environment if known note we only support the most recent release of loris loris version linux distribution and version site hosted on test loris | 1 |
90,157 | 25,995,522,772 | IssuesEvent | 2022-12-20 11:13:19 | niketa282/Calculator | https://api.github.com/repos/niketa282/Calculator | closed | build(wsl): "Add clang format, clang tidy support" | build research | Understand the purpose of these tools , research into any other helpful tools
But remember using too many tools get in the way, choose wisely! | 1.0 | build(wsl): "Add clang format, clang tidy support" - Understand the purpose of these tools , research into any other helpful tools
But remember using too many tools get in the way, choose wisely! | non_test | build wsl add clang format clang tidy support understand the purpose of these tools research into any other helpful tools but remember using too many tools get in the way choose wisely | 0 |
93,856 | 15,946,426,592 | IssuesEvent | 2021-04-15 01:02:31 | jgeraigery/core | https://api.github.com/repos/jgeraigery/core | opened | CVE-2020-35491 (High) detected in jackson-databind-2.9.6.jar | security vulnerability | ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p>
<p>Path to vulnerable library: core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-config-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-core/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.cloud:spring-cloud-starter-config:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-35491","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-35491 (High) detected in jackson-databind-2.9.6.jar - ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p>
<p>Path to vulnerable library: core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-config-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-core/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.cloud:spring-cloud-starter-config:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-35491","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file core nimbus entity dsl pom xml path to vulnerable library core jackson databind jackson databind jar core jackson databind jackson databind jar core jackson databind jackson databind jar dependency hierarchy spring cloud starter config release jar root library x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework cloud spring cloud starter config release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource vulnerabilityurl | 0 |
309,995 | 26,692,246,127 | IssuesEvent | 2023-01-27 06:46:56 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | com.hazelcast.jet.core.SuspendResumeTest.when_cancelSuspendedJob_then_jobCancels | Type: Test-Failure Source: Internal Module: Jet Team: Platform | _master_ (commit 65933d5f0689250b39ad958322ee6fb77baaf936)
Failed on oracle-11: https://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-sonar/1469/testReport/com.hazelcast.jet.core/SuspendResumeTest/when_cancelSuspendedJob_then_jobCancels/
<details><summary>Stacktrace:</summary>
```
java.lang.AssertionError:
'com.hazelcast.jet.impl.exception.CancellationByUserException
at com.hazelcast.jet.impl.MasterJobContext.createCancellationException(MasterJobContext.java:208)
at com.hazelcast.jet.impl.MasterJobContext.requestTermination(MasterJobContext.java:431)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$terminateJob$14(JobCoordinationService.java:494)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$toNullFunction$45(JobCoordinationService.java:945)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$callWithJob$46(JobCoordinationService.java:981)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$submitToCoordinatorThread$55(JobCoordinationService.java:1433)
at com.hazelcast.internal.util.executor.CompletableFutureTask.run(CompletableFutureTask.java:64)
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:217)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102)
' didn't contain expected 'java.util.concurrent.CancellationException'
at org.junit.Assert.fail(Assert.java:89)
at com.hazelcast.test.HazelcastTestSupport.assertContains(HazelcastTestSupport.java:903)
at com.hazelcast.jet.core.SuspendResumeTest.lambda$when_cancelSuspendedJob_then_jobCancels$9(SuspendResumeTest.java:235)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1275)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1379)
at com.hazelcast.jet.core.SuspendResumeTest.when_cancelSuspendedJob_then_jobCancels(SuspendResumeTest.java:232)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.lang.Thread.run(Thread.java:834)
```
</details>
<details><summary>Standard output:</summary>
```
Finished Running Test: when_nonCoordinatorDiesWhileSuspended_then_jobResumes in 2.668 seconds.
Started Running Test: when_cancelSuspendedJob_then_jobCancels
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5701
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,141 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,145 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is STARTING
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:1, ver:1} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
]
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,149 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is STARTED
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5702
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,152 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,155 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,157 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,158 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,158 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is STARTING
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5701-38a78679-68be-4625-8547-2a9e1e3c7817, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=true}
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5702-53de9b30-46a1-4bf5-a8ae-09ba61d3c537, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=true}
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:2} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
]
00:43:32,249 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,260 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:2} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
]
00:43:32,260 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is STARTED
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5703
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,263 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,267 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,269 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,270 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,270 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is STARTING
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5701-38a78679-68be-4625-8547-2a9e1e3c7817, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=true}
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5703-9ff9b136-f005-43c9-84d0-5b098a537c01, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=true}
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
]
00:43:32,272 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5703-9ff9b136-f005-43c9-84d0-5b098a537c01, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=true}
00:43:32,272 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
]
00:43:32,349 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,372 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 this
]
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5702-53de9b30-46a1-4bf5-a8ae-09ba61d3c537, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=true}
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is STARTED
00:43:32,374 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Initializing cluster partition table arrangement...
00:43:32,374 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,374 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,375 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Starting job 0950-6a5b-2540-0001 based on submit request
00:43:32,377 INFO |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Didn't find any snapshot to restore for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 INFO |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Start executing job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001, execution graph in DOT format:
digraph DAG {
"test" [localParallelism=4];
}
HINT: You can use graphviz or http://viz-js.com to visualize the printed graph.
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Building execution plan for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Built execution plans for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - MockPS.init called on hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3
00:43:32,378 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - MockPS.init called on hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - MockPS.init called on hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,379 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Init of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 was successful
00:43:32,379 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Executing job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,380 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [Job] when_cancelSuspendedJob_then_jobCancels - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Sending SUSPEND_GRACEFUL request for job 0950-6a5b-2540-0001 (name ??)
00:43:32,380 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-25 - MockPS.close called on ForkJoinPool.commonPool-worker-25
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:411) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:438) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) ~[classes/:?]
at com.hazelcast.jet.impl.MasterContext.invokeOnParticipant(MasterContext.java:294) ~[classes/:?]
at com.hazelcast.jet.impl.MasterContext.invokeOnParticipants(MasterContext.java:277) ~[classes/:?]
at com.hazelcast.jet.impl.MasterJobContext.lambda$cancelExecutionInvocations$25(MasterJobContext.java:685) ~[classes/:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,381 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-23 - MockPS.close called on ForkJoinPool.commonPool-worker-23
00:43:32,381 DEBUG || - [MasterJobContext] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5701: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,381 DEBUG || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-27 - MockPS.close called on ForkJoinPool.commonPool-worker-27
00:43:32,381 DEBUG || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:498) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:498) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5703: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5702: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 has failures: [[127.0.0.1]:5701=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL, [127.0.0.1]:5703=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL, [127.0.0.1]:5702=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL]
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,451 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:32,581 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [Job] when_cancelSuspendedJob_then_jobCancels - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Sending CANCEL_FORCEFUL request for job 0950-6a5b-2540-0001 (name ??)
00:43:32,581 WARN |when_cancelSuspendedJob_then_jobCancels| - [JoinSubmittedJobOperation] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] null
com.hazelcast.jet.impl.exception.CancellationByUserException: null
at com.hazelcast.jet.impl.MasterJobContext.createCancellationException(MasterJobContext.java:208) ~[classes/:?]
at com.hazelcast.jet.impl.MasterJobContext.requestTermination(MasterJobContext.java:416) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$terminateJob$14(JobCoordinationService.java:494) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$toNullFunction$45(JobCoordinationService.java:945) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$callWithJob$46(JobCoordinationService.java:981) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$submitToCoordinatorThread$55(JobCoordinationService.java:1433) ~[classes/:?]
at com.hazelcast.internal.util.executor.CompletableFutureTask.run(CompletableFutureTask.java:64) ~[classes/:?]
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:217) ~[classes/:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,582 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 is completed
00:43:37,451 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:42,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:47,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:52,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:57,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:02,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-6 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:07,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-12 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:12,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:17,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-5 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:22,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-13 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:27,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:32,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-1 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:37,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-1 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:42,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:47,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:52,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-10 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:57,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-5 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:02,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-12 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:07,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-9 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 1ms
00:45:12,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:17,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-14 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:22,458 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-9 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:27,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 1ms
00:45:32,458 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:34,914 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetTestSupport] Time-limited test - Terminating instanceFactory in JetTestSupport.@After
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is SHUTTING_DOWN
00:45:34,915 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=false}
00:45:34,915 WARN |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 is suspected to be dead for reason: Connection manager is stopped on Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 this
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removing Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:4} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
]
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5703, UUID: 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_silly_chebyshev.cached.thread-2 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5703, UUID: 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:4} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
]
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is SHUTDOWN
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is SHUTTING_DOWN
00:45:34,918 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=false}
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=false}
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removing Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MigrationManager] hz.SuspendResumeTest_trusting_chebyshev.migration - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Repartitioning cluster data. Migration tasks count: 7
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:1, ver:5} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
]
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-6 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5702, UUID: 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MigrationManager] hz.SuspendResumeTest_trusting_chebyshev.migration - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Rebalance process was failed. Ignoring remaining migrations. Will recalculate the new migration plan. (repartitionTime=Fri Jan 27 00:45:34 UTC 2023, plannedMigrations=7, completedMigrations=0, remainingMigrations=7, totalCompletedMigrations=0, elapsedMigrationOperationTime=0ms, totalElapsedMigrationOperationTime=0ms, elapsedDestinationCommitTime=0ms, totalElapsedDestinationCommitTime=0ms, elapsedMigrationTime=0ms, totalElapsedMigrationTime=0ms)
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is SHUTDOWN
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is SHUTTING_DOWN
00:45:34,921 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,923 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,924 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,924 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is SHUTDOWN
BuildInfo right after when_cancelSuspendedJob_then_jobCancels(com.hazelcast.jet.core.SuspendResumeTest): BuildInfo{version='5.3.0-SNAPSHOT', build='20230126', buildNumber=20230126, revision=65933d5, enterprise=false, serializationVersion=1}
Hiccups measured while running test 'when_cancelSuspendedJob_then_jobCancels(com.hazelcast.jet.core.SuspendResumeTest):'
00:43:30, accumulated pauses: 52 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:35, accumulated pauses: 54 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:40, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:45, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:50, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:55, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:00, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:05, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:10, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:15, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:20, accumulated pauses: 57 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:25, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:30, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:35, accumulated pauses: 57 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:40, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:45, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:50, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:55, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:00, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:05, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:10, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:15, accumulated pauses: 61 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:20, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:25, accumulated pauses: 64 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:30, accumulated pauses: 2190 ms, max pause: 2155 ms, pauses over 1000 ms: 1
No metrics recorded during the test
```
</details>
Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-master-sonar/1469/
| 1.0 | com.hazelcast.jet.core.SuspendResumeTest.when_cancelSuspendedJob_then_jobCancels - _master_ (commit 65933d5f0689250b39ad958322ee6fb77baaf936)
Failed on oracle-11: https://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-sonar/1469/testReport/com.hazelcast.jet.core/SuspendResumeTest/when_cancelSuspendedJob_then_jobCancels/
<details><summary>Stacktrace:</summary>
```
java.lang.AssertionError:
'com.hazelcast.jet.impl.exception.CancellationByUserException
at com.hazelcast.jet.impl.MasterJobContext.createCancellationException(MasterJobContext.java:208)
at com.hazelcast.jet.impl.MasterJobContext.requestTermination(MasterJobContext.java:431)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$terminateJob$14(JobCoordinationService.java:494)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$toNullFunction$45(JobCoordinationService.java:945)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$callWithJob$46(JobCoordinationService.java:981)
at com.hazelcast.jet.impl.JobCoordinationService.lambda$submitToCoordinatorThread$55(JobCoordinationService.java:1433)
at com.hazelcast.internal.util.executor.CompletableFutureTask.run(CompletableFutureTask.java:64)
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:217)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102)
' didn't contain expected 'java.util.concurrent.CancellationException'
at org.junit.Assert.fail(Assert.java:89)
at com.hazelcast.test.HazelcastTestSupport.assertContains(HazelcastTestSupport.java:903)
at com.hazelcast.jet.core.SuspendResumeTest.lambda$when_cancelSuspendedJob_then_jobCancels$9(SuspendResumeTest.java:235)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1275)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1379)
at com.hazelcast.jet.core.SuspendResumeTest.when_cancelSuspendedJob_then_jobCancels(SuspendResumeTest.java:232)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.lang.Thread.run(Thread.java:834)
```
</details>
<details><summary>Standard output:</summary>
```
Finished Running Test: when_nonCoordinatorDiesWhileSuspended_then_jobResumes in 2.668 seconds.
Started Running Test: when_cancelSuspendedJob_then_jobCancels
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5701
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,138 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,141 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,145 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,148 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is STARTING
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:1, ver:1} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
]
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,149 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is STARTED
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5702
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,149 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,152 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,155 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,157 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,158 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,158 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is STARTING
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5701-38a78679-68be-4625-8547-2a9e1e3c7817, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=true}
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5702-53de9b30-46a1-4bf5-a8ae-09ba61d3c537, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=true}
00:43:32,159 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:2} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
]
00:43:32,249 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,260 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:2} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
]
00:43:32,260 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is STARTED
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [LOCAL] [dev] [5.3.0-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1'
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [logo] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT]
+ + o o o o---o o----o o o---o o o----o o--o--o
+ + + + | | / \ / | | / / \ | |
+ + + + + o----o o o o o----o | o o o o----o |
+ + + + | | / \ / | | \ / \ | |
+ + o o o o o---o o----o o----o o---o o o o----o o
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Copyright (c) 2008-2023, Hazelcast, Inc. All Rights Reserved.
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Hazelcast Platform 5.3.0-SNAPSHOT (20230126 - 65933d5) starting at [127.0.0.1]:5703
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Cluster name: dev
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker.
00:43:32,261 INFO |when_cancelSuspendedJob_then_jobCancels| - [system] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Jet is enabled
00:43:32,263 INFO |when_cancelSuspendedJob_then_jobCancels| - [MetricsConfigHelper] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled
00:43:32,267 WARN |when_cancelSuspendedJob_then_jobCancels| - [CPSubsystem] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees.
00:43:32,269 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetServiceBackend] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Setting number of cooperative threads and default parallelism to 4
00:43:32,270 INFO |when_cancelSuspendedJob_then_jobCancels| - [Diagnostics] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments.
00:43:32,270 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is STARTING
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5701-38a78679-68be-4625-8547-2a9e1e3c7817, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=true}
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5703-9ff9b136-f005-43c9-84d0-5b098a537c01, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=true}
00:43:32,271 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_trusting_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
]
00:43:32,272 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5703-9ff9b136-f005-43c9-84d0-5b098a537c01, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=true}
00:43:32,272 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
]
00:43:32,349 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Not starting jobs because partitions are not yet initialized.
00:43:32,372 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT]
Members {size:3, ver:3} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 this
]
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Created connection to endpoint: [127.0.0.1]:5702-53de9b30-46a1-4bf5-a8ae-09ba61d3c537, connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=true}
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Jet started scanning for jobs
00:43:32,373 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Time-limited test - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is STARTED
00:43:32,374 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Initializing cluster partition table arrangement...
00:43:32,374 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,374 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,375 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Starting job 0950-6a5b-2540-0001 based on submit request
00:43:32,377 INFO |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Didn't find any snapshot to restore for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 INFO |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Start executing job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001, execution graph in DOT format:
digraph DAG {
"test" [localParallelism=4];
}
HINT: You can use graphviz or http://viz-js.com to visualize the printed graph.
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Building execution plan for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Built execution plans for job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,377 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - MockPS.init called on hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3
00:43:32,378 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [InitExecutionOperation] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Initializing execution plan for job 0950-6a5b-2540-0001, execution 0950-6a5b-2541-0001 from [127.0.0.1]:5701
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Creating job classLoader for job 0950-6a5b-2540-0001
00:43:32,378 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Create processor classloader map for job 0950-6a5b-2540-0001
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - MockPS.init called on hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [HazelcastTestSupport] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - MockPS.init called on hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-1 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Execution plan for jobId=0950-6a5b-2540-0001, jobName='0950-6a5b-2540-0001', executionId=0950-6a5b-2541-0001 initialized
00:43:32,379 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Init of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 was successful
00:43:32,379 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Executing job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_hardcore_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,379 INFO |when_cancelSuspendedJob_then_jobCancels| - [JobExecutionService] hz.SuspendResumeTest_silly_chebyshev.generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Start execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 from coordinator [127.0.0.1]:5701
00:43:32,380 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [Job] when_cancelSuspendedJob_then_jobCancels - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Sending SUSPEND_GRACEFUL request for job 0950-6a5b-2540-0001 (name ??)
00:43:32,380 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-25 - MockPS.close called on ForkJoinPool.commonPool-worker-25
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:411) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:438) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) ~[classes/:?]
at com.hazelcast.jet.impl.MasterContext.invokeOnParticipant(MasterContext.java:294) ~[classes/:?]
at com.hazelcast.jet.impl.MasterContext.invokeOnParticipants(MasterContext.java:277) ~[classes/:?]
at com.hazelcast.jet.impl.MasterJobContext.lambda$cancelExecutionInvocations$25(MasterJobContext.java:685) ~[classes/:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,381 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-23 - MockPS.close called on ForkJoinPool.commonPool-worker-23
00:43:32,381 DEBUG || - [MasterJobContext] ForkJoinPool.commonPool-worker-25 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5701: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,381 DEBUG || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 INFO || - [HazelcastTestSupport] ForkJoinPool.commonPool-worker-27 - MockPS.close called on ForkJoinPool.commonPool-worker-27
00:43:32,381 DEBUG || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Completed execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-23 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:498) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,381 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 completed with failure
java.util.concurrent.CompletionException: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?]
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?]
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?]
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?]
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?]
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?]
at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:498) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:429) ~[classes/:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:415) ~[classes/:?]
at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:807) ~[?:?]
at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:369) ~[classes/:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
Caused by: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
at com.hazelcast.jet.impl.execution.ExecutionContext.terminateExecution(ExecutionContext.java:318) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution0(JobExecutionService.java:722) ~[classes/:?]
at com.hazelcast.jet.impl.JobExecutionService.terminateExecution(JobExecutionService.java:718) ~[classes/:?]
at com.hazelcast.jet.impl.operation.TerminateExecutionOperation.doRun(TerminateExecutionOperation.java:59) ~[classes/:?]
at com.hazelcast.jet.impl.operation.AsyncOperation.run(AsyncOperation.java:55) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:295) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) ~[classes/:?]
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:498) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) ~[classes/:?]
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5703: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 received response to StartExecutionOperation from [127.0.0.1]:5702: com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [MasterJobContext] ForkJoinPool.commonPool-worker-27 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Execution of job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 has failures: [[127.0.0.1]:5701=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL, [127.0.0.1]:5703=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL, [127.0.0.1]:5702=com.hazelcast.jet.impl.exception.JobTerminateRequestedException: SUSPEND_FORCEFUL]
00:43:32,382 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobClassLoaderService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Finish JobClassLoaders phaseCount = 0, removing classloaders for jobId=0950-6a5b-2540-0001
00:43:32,451 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:32,581 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [Job] when_cancelSuspendedJob_then_jobCancels - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Sending CANCEL_FORCEFUL request for job 0950-6a5b-2540-0001 (name ??)
00:43:32,581 WARN |when_cancelSuspendedJob_then_jobCancels| - [JoinSubmittedJobOperation] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] null
com.hazelcast.jet.impl.exception.CancellationByUserException: null
at com.hazelcast.jet.impl.MasterJobContext.createCancellationException(MasterJobContext.java:208) ~[classes/:?]
at com.hazelcast.jet.impl.MasterJobContext.requestTermination(MasterJobContext.java:416) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$terminateJob$14(JobCoordinationService.java:494) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$toNullFunction$45(JobCoordinationService.java:945) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$callWithJob$46(JobCoordinationService.java:981) ~[classes/:?]
at com.hazelcast.jet.impl.JobCoordinationService.lambda$submitToCoordinatorThread$55(JobCoordinationService.java:1433) ~[classes/:?]
at com.hazelcast.internal.util.executor.CompletableFutureTask.run(CompletableFutureTask.java:64) ~[classes/:?]
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:217) ~[classes/:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[classes/:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[classes/:?]
00:43:32,582 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobCoordinationService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] job '0950-6a5b-2540-0001', execution 0950-6a5b-2541-0001 is completed
00:43:37,451 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:42,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:47,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-8 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:52,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:43:57,453 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:02,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-6 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:07,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-12 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:12,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:17,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-5 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:22,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-13 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:27,454 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:32,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-1 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:37,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-1 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:42,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:47,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:52,455 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-10 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:44:57,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-5 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:02,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-12 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:07,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-9 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 1ms
00:45:12,456 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:17,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-14 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:22,458 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-9 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:27,457 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-7 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 1ms
00:45:32,458 DEBUG |when_cancelSuspendedJob_then_jobCancels| - [JobRepository] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Job cleanup took 0ms
00:45:34,914 INFO |when_cancelSuspendedJob_then_jobCancels| - [JetTestSupport] Time-limited test - Terminating instanceFactory in JetTestSupport.@After
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is SHUTTING_DOWN
00:45:34,915 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], alive=false}
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5703, uuid=9ff9b136-f005-43c9-84d0-5b098a537c01], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=false}
00:45:34,915 WARN |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 is suspected to be dead for reason: Connection manager is stopped on Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01 this
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removing Member [127.0.0.1]:5703 - 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,915 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:4} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
]
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-4 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5703, UUID: 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_silly_chebyshev.cached.thread-2 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5703, UUID: 9ff9b136-f005-43c9-84d0-5b098a537c01
00:45:34,916 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] hz.SuspendResumeTest_silly_chebyshev.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT]
Members {size:2, ver:4} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817
Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537 this
]
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5703 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5703 is SHUTDOWN
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is SHUTTING_DOWN
00:45:34,918 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], remoteEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], alive=false}
00:45:34,918 INFO |when_cancelSuspendedJob_then_jobCancels| - [MockServer] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Removed connection to endpoint: [address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], connection: MockConnection{localEndpoint=[address=[127.0.0.1]:5702, uuid=53de9b30-46a1-4bf5-a8ae-09ba61d3c537], remoteEndpoint=[address=[127.0.0.1]:5701, uuid=38a78679-68be-4625-8547-2a9e1e3c7817], alive=false}
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MembershipManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Removing Member [127.0.0.1]:5702 - 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MigrationManager] hz.SuspendResumeTest_trusting_chebyshev.migration - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Repartitioning cluster data. Migration tasks count: 7
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [PartitionStateManager] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Storing snapshot of partition assignments while removing UUID 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [ClusterService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT]
Members {size:1, ver:5} [
Member [127.0.0.1]:5701 - 38a78679-68be-4625-8547-2a9e1e3c7817 this
]
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [TransactionManagerService] hz.SuspendResumeTest_trusting_chebyshev.cached.thread-6 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Committing/rolling-back live transactions of [127.0.0.1]:5702, UUID: 53de9b30-46a1-4bf5-a8ae-09ba61d3c537
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,919 INFO |when_cancelSuspendedJob_then_jobCancels| - [MigrationManager] hz.SuspendResumeTest_trusting_chebyshev.migration - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Rebalance process was failed. Ignoring remaining migrations. Will recalculate the new migration plan. (repartitionTime=Fri Jan 27 00:45:34 UTC 2023, plannedMigrations=7, completedMigrations=0, remainingMigrations=7, totalCompletedMigrations=0, elapsedMigrationOperationTime=0ms, totalElapsedMigrationOperationTime=0ms, elapsedDestinationCommitTime=0ms, totalElapsedDestinationCommitTime=0ms, elapsedMigrationTime=0ms, totalElapsedMigrationTime=0ms)
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5702 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5702 is SHUTDOWN
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is SHUTTING_DOWN
00:45:34,921 WARN |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Terminating forcefully...
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Shutting down connection manager...
00:45:34,921 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Shutting down node engine...
00:45:34,923 INFO |when_cancelSuspendedJob_then_jobCancels| - [NodeExtension] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Destroying node NodeExtension.
00:45:34,924 INFO |when_cancelSuspendedJob_then_jobCancels| - [Node] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] Hazelcast Shutdown is completed in 3 ms.
00:45:34,924 INFO |when_cancelSuspendedJob_then_jobCancels| - [LifecycleService] Thread-56518 - [127.0.0.1]:5701 [dev] [5.3.0-SNAPSHOT] [127.0.0.1]:5701 is SHUTDOWN
BuildInfo right after when_cancelSuspendedJob_then_jobCancels(com.hazelcast.jet.core.SuspendResumeTest): BuildInfo{version='5.3.0-SNAPSHOT', build='20230126', buildNumber=20230126, revision=65933d5, enterprise=false, serializationVersion=1}
Hiccups measured while running test 'when_cancelSuspendedJob_then_jobCancels(com.hazelcast.jet.core.SuspendResumeTest):'
00:43:30, accumulated pauses: 52 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:35, accumulated pauses: 54 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:40, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:45, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:50, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:43:55, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:00, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:05, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:10, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:15, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:20, accumulated pauses: 57 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:25, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:30, accumulated pauses: 55 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:35, accumulated pauses: 57 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:40, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:45, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:50, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:44:55, accumulated pauses: 56 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:00, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:05, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:10, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:15, accumulated pauses: 61 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:20, accumulated pauses: 63 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:25, accumulated pauses: 64 ms, max pause: 0 ms, pauses over 1000 ms: 0
00:45:30, accumulated pauses: 2190 ms, max pause: 2155 ms, pauses over 1000 ms: 1
No metrics recorded during the test
```
</details>
Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-master-sonar/1469/
| test | com hazelcast jet core suspendresumetest when cancelsuspendedjob then jobcancels master commit failed on oracle stacktrace java lang assertionerror com hazelcast jet impl exception cancellationbyuserexception at com hazelcast jet impl masterjobcontext createcancellationexception masterjobcontext java at com hazelcast jet impl masterjobcontext requesttermination masterjobcontext java at com hazelcast jet impl jobcoordinationservice lambda terminatejob jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda tonullfunction jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda callwithjob jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda submittocoordinatorthread jobcoordinationservice java at com hazelcast internal util executor completablefuturetask run completablefuturetask java at com hazelcast internal util executor cachedexecutorservicedelegate worker run cachedexecutorservicedelegate java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java didn t contain expected java util concurrent cancellationexception at org junit assert fail assert java at com hazelcast test hazelcasttestsupport assertcontains hazelcasttestsupport java at com hazelcast jet core suspendresumetest lambda when cancelsuspendedjob then jobcancels suspendresumetest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast jet core suspendresumetest when cancelsuspendedjob then jobcancels suspendresumetest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java base java util concurrent futuretask run futuretask java at java base java lang thread run thread java standard output finished running test when noncoordinatordieswhilesuspended then jobresumes in seconds started running test when cancelsuspendedjob then jobcancels info when cancelsuspendedjob then jobcancels time limited test overridden metrics configuration with system property hazelcast metrics collection frequency metricsconfig collectionfrequencyseconds info when cancelsuspendedjob then jobcancels time limited test o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o info when cancelsuspendedjob then jobcancels time limited test copyright c hazelcast inc all rights reserved info when cancelsuspendedjob then jobcancels time limited test hazelcast platform snapshot starting at info when cancelsuspendedjob then jobcancels time limited test cluster name dev info when cancelsuspendedjob then jobcancels time limited test integrity checker is disabled fail fast on corrupted executables will not be performed for more information see the documentation for integrity checker info when cancelsuspendedjob then jobcancels time limited test jet is enabled info when cancelsuspendedjob then jobcancels time limited test collecting debug metrics and sending to diagnostics is enabled warn when cancelsuspendedjob then jobcancels time limited test cp subsystem is not enabled cp data structures will operate in unsafe mode please note that unsafe mode will not provide strong consistency guarantees info when cancelsuspendedjob then jobcancels time limited test setting number of cooperative threads and default parallelism to info when cancelsuspendedjob then jobcancels time limited test diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments info when cancelsuspendedjob then jobcancels time limited test is starting info when cancelsuspendedjob then jobcancels time limited test members size ver member this info when cancelsuspendedjob then jobcancels time limited test jet started scanning for jobs debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread not starting jobs because partitions are not yet initialized info when cancelsuspendedjob then jobcancels time limited test is started info when cancelsuspendedjob then jobcancels time limited test overridden metrics configuration with system property hazelcast metrics collection frequency metricsconfig collectionfrequencyseconds info when cancelsuspendedjob then jobcancels time limited test o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o info when cancelsuspendedjob then jobcancels time limited test copyright c hazelcast inc all rights reserved info when cancelsuspendedjob then jobcancels time limited test hazelcast platform snapshot starting at info when cancelsuspendedjob then jobcancels time limited test cluster name dev info when cancelsuspendedjob then jobcancels time limited test integrity checker is disabled fail fast on corrupted executables will not be performed for more information see the documentation for integrity checker info when cancelsuspendedjob then jobcancels time limited test jet is enabled info when cancelsuspendedjob then jobcancels time limited test collecting debug metrics and sending to diagnostics is enabled warn when cancelsuspendedjob then jobcancels time limited test cp subsystem is not enabled cp data structures will operate in unsafe mode please note that unsafe mode will not provide strong consistency guarantees info when cancelsuspendedjob then jobcancels time limited test setting number of cooperative threads and default parallelism to info when cancelsuspendedjob then jobcancels time limited test diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments info when cancelsuspendedjob then jobcancels time limited test is starting info when cancelsuspendedjob then jobcancels time limited test created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev priority generic operation thread created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev priority generic operation thread members size ver member this member debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread not starting jobs because partitions are not yet initialized info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev priority generic operation thread members size ver member member this info when cancelsuspendedjob then jobcancels time limited test jet started scanning for jobs info when cancelsuspendedjob then jobcancels time limited test is started info when cancelsuspendedjob then jobcancels time limited test overridden metrics configuration with system property hazelcast metrics collection frequency metricsconfig collectionfrequencyseconds info when cancelsuspendedjob then jobcancels time limited test o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o info when cancelsuspendedjob then jobcancels time limited test copyright c hazelcast inc all rights reserved info when cancelsuspendedjob then jobcancels time limited test hazelcast platform snapshot starting at info when cancelsuspendedjob then jobcancels time limited test cluster name dev info when cancelsuspendedjob then jobcancels time limited test integrity checker is disabled fail fast on corrupted executables will not be performed for more information see the documentation for integrity checker info when cancelsuspendedjob then jobcancels time limited test jet is enabled info when cancelsuspendedjob then jobcancels time limited test collecting debug metrics and sending to diagnostics is enabled warn when cancelsuspendedjob then jobcancels time limited test cp subsystem is not enabled cp data structures will operate in unsafe mode please note that unsafe mode will not provide strong consistency guarantees info when cancelsuspendedjob then jobcancels time limited test setting number of cooperative threads and default parallelism to info when cancelsuspendedjob then jobcancels time limited test diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments info when cancelsuspendedjob then jobcancels time limited test is starting info when cancelsuspendedjob then jobcancels time limited test created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev priority generic operation thread created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev priority generic operation thread members size ver member this member member info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev priority generic operation thread created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev priority generic operation thread members size ver member member this member debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread not starting jobs because partitions are not yet initialized info when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread members size ver member member member this info when cancelsuspendedjob then jobcancels time limited test created connection to endpoint connection mockconnection localendpoint uuid remoteendpoint uuid alive true info when cancelsuspendedjob then jobcancels time limited test jet started scanning for jobs info when cancelsuspendedjob then jobcancels time limited test is started info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread initializing cluster partition table arrangement debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread creating job classloader for job debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread create processor classloader map for job info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread starting job based on submit request info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread didn t find any snapshot to restore for job execution info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread start executing job execution execution graph in dot format digraph dag test hint you can use graphviz or to visualize the printed graph debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread building execution plan for job execution debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread built execution plans for job execution debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread initializing execution plan for job execution from info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread mockps init called on hz suspendresumetest trusting chebyshev cached thread info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread execution plan for jobid jobname executionid initialized debug when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread initializing execution plan for job execution from debug when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread initializing execution plan for job execution from debug when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread creating job classloader for job debug when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread create processor classloader map for job debug when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread creating job classloader for job debug when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread create processor classloader map for job info when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread mockps init called on hz suspendresumetest hardcore chebyshev generic operation thread info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread mockps init called on hz suspendresumetest silly chebyshev generic operation thread info when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread execution plan for jobid jobname executionid initialized info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread execution plan for jobid jobname executionid initialized debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread init of job execution was successful debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread executing job execution info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread start execution of job execution from coordinator info when cancelsuspendedjob then jobcancels hz suspendresumetest hardcore chebyshev generic operation thread start execution of job execution from coordinator info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev generic operation thread start execution of job execution from coordinator debug when cancelsuspendedjob then jobcancels when cancelsuspendedjob then jobcancels sending suspend graceful request for job name info forkjoinpool commonpool worker mockps close called on forkjoinpool commonpool worker debug forkjoinpool commonpool worker completed execution of job execution debug forkjoinpool commonpool worker execution of job execution completed with failure java util concurrent completionexception com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at java util concurrent completablefuture encodethrowable completablefuture java at java util concurrent completablefuture completethrowable completablefuture java at java util concurrent completablefuture uniwhencomplete completablefuture java at java util concurrent completablefuture uniwhencomplete tryfire completablefuture java at java util concurrent completablefuture postcomplete completablefuture java at java util concurrent completablefuture completeexceptionally completablefuture java at com hazelcast jet impl util noncompletablefuture internalcompleteexceptionally noncompletablefuture java at com hazelcast jet impl execution taskletexecutionservice executiontracker taskletdone taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker dismisstasklet taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java at java util concurrent copyonwritearraylist foreach copyonwritearraylist java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker run taskletexecutionservice java at java lang thread run thread java caused by com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at com hazelcast jet impl execution executioncontext terminateexecution executioncontext java at com hazelcast jet impl jobexecutionservice jobexecutionservice java at com hazelcast jet impl jobexecutionservice terminateexecution jobexecutionservice java at com hazelcast jet impl operation terminateexecutionoperation dorun terminateexecutionoperation java at com hazelcast jet impl operation asyncoperation run asyncoperation java at com hazelcast spi impl operationservice operation call operation java at com hazelcast spi impl operationservice impl operationrunnerimpl call operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl run operationexecutorimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl runorexecute operationexecutorimpl java at com hazelcast spi impl operationservice impl invocation doinvokelocal invocation java at com hazelcast spi impl operationservice impl invocation doinvoke invocation java at com hazelcast spi impl operationservice impl invocation invocation java at com hazelcast spi impl operationservice impl invocation invoke invocation java at com hazelcast spi impl operationservice impl invocationbuilderimpl invoke invocationbuilderimpl java at com hazelcast jet impl mastercontext invokeonparticipant mastercontext java at com hazelcast jet impl mastercontext invokeonparticipants mastercontext java at com hazelcast jet impl masterjobcontext lambda cancelexecutioninvocations masterjobcontext java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java info forkjoinpool commonpool worker mockps close called on forkjoinpool commonpool worker debug forkjoinpool commonpool worker job execution received response to startexecutionoperation from com hazelcast jet impl exception jobterminaterequestedexception suspend forceful debug forkjoinpool commonpool worker finish jobclassloaders phasecount removing classloaders for jobid debug forkjoinpool commonpool worker completed execution of job execution info forkjoinpool commonpool worker mockps close called on forkjoinpool commonpool worker debug forkjoinpool commonpool worker finish jobclassloaders phasecount removing classloaders for jobid debug forkjoinpool commonpool worker completed execution of job execution debug forkjoinpool commonpool worker execution of job execution completed with failure java util concurrent completionexception com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at java util concurrent completablefuture encodethrowable completablefuture java at java util concurrent completablefuture completethrowable completablefuture java at java util concurrent completablefuture uniwhencomplete completablefuture java at java util concurrent completablefuture uniwhencomplete tryfire completablefuture java at java util concurrent completablefuture postcomplete completablefuture java at java util concurrent completablefuture completeexceptionally completablefuture java at com hazelcast jet impl util noncompletablefuture internalcompleteexceptionally noncompletablefuture java at com hazelcast jet impl execution taskletexecutionservice executiontracker taskletdone taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker dismisstasklet taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java at java util concurrent copyonwritearraylist foreach copyonwritearraylist java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker run taskletexecutionservice java at java lang thread run thread java caused by com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at com hazelcast jet impl execution executioncontext terminateexecution executioncontext java at com hazelcast jet impl jobexecutionservice jobexecutionservice java at com hazelcast jet impl jobexecutionservice terminateexecution jobexecutionservice java at com hazelcast jet impl operation terminateexecutionoperation dorun terminateexecutionoperation java at com hazelcast jet impl operation asyncoperation run asyncoperation java at com hazelcast spi impl operationservice operation call operation java at com hazelcast spi impl operationservice impl operationrunnerimpl call operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread executerun operationthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java debug forkjoinpool commonpool worker execution of job execution completed with failure java util concurrent completionexception com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at java util concurrent completablefuture encodethrowable completablefuture java at java util concurrent completablefuture completethrowable completablefuture java at java util concurrent completablefuture uniwhencomplete completablefuture java at java util concurrent completablefuture uniwhencomplete tryfire completablefuture java at java util concurrent completablefuture postcomplete completablefuture java at java util concurrent completablefuture completeexceptionally completablefuture java at com hazelcast jet impl util noncompletablefuture internalcompleteexceptionally noncompletablefuture java at com hazelcast jet impl execution taskletexecutionservice executiontracker taskletdone taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker dismisstasklet taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java at java util concurrent copyonwritearraylist foreach copyonwritearraylist java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker run taskletexecutionservice java at java lang thread run thread java caused by com hazelcast jet impl exception jobterminaterequestedexception suspend forceful at com hazelcast jet impl execution executioncontext terminateexecution executioncontext java at com hazelcast jet impl jobexecutionservice jobexecutionservice java at com hazelcast jet impl jobexecutionservice terminateexecution jobexecutionservice java at com hazelcast jet impl operation terminateexecutionoperation dorun terminateexecutionoperation java at com hazelcast jet impl operation asyncoperation run asyncoperation java at com hazelcast spi impl operationservice operation call operation java at com hazelcast spi impl operationservice impl operationrunnerimpl call operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread executerun operationthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java debug when cancelsuspendedjob then jobcancels forkjoinpool commonpool worker job execution received response to startexecutionoperation from com hazelcast jet impl exception jobterminaterequestedexception suspend forceful debug when cancelsuspendedjob then jobcancels forkjoinpool commonpool worker job execution received response to startexecutionoperation from com hazelcast jet impl exception jobterminaterequestedexception suspend forceful debug when cancelsuspendedjob then jobcancels forkjoinpool commonpool worker execution of job execution has failures com hazelcast jet impl exception jobterminaterequestedexception suspend forceful com hazelcast jet impl exception jobterminaterequestedexception suspend forceful com hazelcast jet impl exception jobterminaterequestedexception suspend forceful debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread finish jobclassloaders phasecount removing classloaders for jobid debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels when cancelsuspendedjob then jobcancels sending cancel forceful request for job name warn when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread null com hazelcast jet impl exception cancellationbyuserexception null at com hazelcast jet impl masterjobcontext createcancellationexception masterjobcontext java at com hazelcast jet impl masterjobcontext requesttermination masterjobcontext java at com hazelcast jet impl jobcoordinationservice lambda terminatejob jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda tonullfunction jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda callwithjob jobcoordinationservice java at com hazelcast jet impl jobcoordinationservice lambda submittocoordinatorthread jobcoordinationservice java at com hazelcast internal util executor completablefuturetask run completablefuturetask java at com hazelcast internal util executor cachedexecutorservicedelegate worker run cachedexecutorservicedelegate java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job execution is completed debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took debug when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread job cleanup took info when cancelsuspendedjob then jobcancels time limited test terminating instancefactory in jettestsupport after info when cancelsuspendedjob then jobcancels thread is shutting down warn when cancelsuspendedjob then jobcancels thread terminating forcefully info when cancelsuspendedjob then jobcancels thread shutting down connection manager info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false warn when cancelsuspendedjob then jobcancels thread member is suspected to be dead for reason connection manager is stopped on member this info when cancelsuspendedjob then jobcancels thread removing member info when cancelsuspendedjob then jobcancels thread storing snapshot of partition assignments while removing uuid info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev priority generic operation thread storing snapshot of partition assignments while removing uuid info when cancelsuspendedjob then jobcancels thread members size ver member this member info when cancelsuspendedjob then jobcancels thread shutting down node engine info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread committing rolling back live transactions of uuid info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev cached thread committing rolling back live transactions of uuid info when cancelsuspendedjob then jobcancels hz suspendresumetest silly chebyshev priority generic operation thread members size ver member member this info when cancelsuspendedjob then jobcancels thread destroying node nodeextension info when cancelsuspendedjob then jobcancels thread hazelcast shutdown is completed in ms info when cancelsuspendedjob then jobcancels thread is shutdown info when cancelsuspendedjob then jobcancels thread is shutting down warn when cancelsuspendedjob then jobcancels thread terminating forcefully info when cancelsuspendedjob then jobcancels thread shutting down connection manager info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false info when cancelsuspendedjob then jobcancels thread removed connection to endpoint uuid connection mockconnection localendpoint uuid remoteendpoint uuid alive false info when cancelsuspendedjob then jobcancels thread removing member info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev migration repartitioning cluster data migration tasks count info when cancelsuspendedjob then jobcancels thread storing snapshot of partition assignments while removing uuid info when cancelsuspendedjob then jobcancels thread members size ver member this info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev cached thread committing rolling back live transactions of uuid info when cancelsuspendedjob then jobcancels thread shutting down node engine info when cancelsuspendedjob then jobcancels hz suspendresumetest trusting chebyshev migration rebalance process was failed ignoring remaining migrations will recalculate the new migration plan repartitiontime fri jan utc plannedmigrations completedmigrations remainingmigrations totalcompletedmigrations elapsedmigrationoperationtime totalelapsedmigrationoperationtime elapseddestinationcommittime totalelapseddestinationcommittime elapsedmigrationtime totalelapsedmigrationtime info when cancelsuspendedjob then jobcancels thread destroying node nodeextension info when cancelsuspendedjob then jobcancels thread hazelcast shutdown is completed in ms info when cancelsuspendedjob then jobcancels thread is shutdown info when cancelsuspendedjob then jobcancels thread is shutting down warn when cancelsuspendedjob then jobcancels thread terminating forcefully info when cancelsuspendedjob then jobcancels thread shutting down connection manager info when cancelsuspendedjob then jobcancels thread shutting down node engine info when cancelsuspendedjob then jobcancels thread destroying node nodeextension info when cancelsuspendedjob then jobcancels thread hazelcast shutdown is completed in ms info when cancelsuspendedjob then jobcancels thread is shutdown buildinfo right after when cancelsuspendedjob then jobcancels com hazelcast jet core suspendresumetest buildinfo version snapshot build buildnumber revision enterprise false serializationversion hiccups measured while running test when cancelsuspendedjob then jobcancels com hazelcast jet core suspendresumetest accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms accumulated pauses ms max pause ms pauses over ms no metrics recorded during the test standard output can be found here | 1 |
291,749 | 8,949,248,773 | IssuesEvent | 2019-01-25 06:46:24 | Lukas-VanDyke/Software-Testing-a1 | https://api.github.com/repos/Lukas-VanDyke/Software-Testing-a1 | closed | [Reopened] System deposits wrong amount | High Priority High Severity | V1.1 Outcome: No longer subtracts $10. Now subtracts $0.10. Total is $1004.90
Function Tested: Deposit to savings
Initial State of System: Off
Steps to reproduce:
1. Press on
2. Enter 1 for number of $20 bills
3. Press insert card
4. Enter correct card number and PIN
5. Select deposit
6. Select savings
7. Enter $5.00 to deposit
8. Click to insert envelope
Expected Outcome: Account total will be $1005.00
Actual Outcome: Account total will be $995.00.The atm will subtract $10 from the savings account after depositing any amount.
| 1.0 | [Reopened] System deposits wrong amount - V1.1 Outcome: No longer subtracts $10. Now subtracts $0.10. Total is $1004.90
Function Tested: Deposit to savings
Initial State of System: Off
Steps to reproduce:
1. Press on
2. Enter 1 for number of $20 bills
3. Press insert card
4. Enter correct card number and PIN
5. Select deposit
6. Select savings
7. Enter $5.00 to deposit
8. Click to insert envelope
Expected Outcome: Account total will be $1005.00
Actual Outcome: Account total will be $995.00.The atm will subtract $10 from the savings account after depositing any amount.
| non_test | system deposits wrong amount outcome no longer subtracts now subtracts total is function tested deposit to savings initial state of system off steps to reproduce press on enter for number of bills press insert card enter correct card number and pin select deposit select savings enter to deposit click to insert envelope expected outcome account total will be actual outcome account total will be the atm will subtract from the savings account after depositing any amount | 0 |
11,207 | 7,468,265,902 | IssuesEvent | 2018-04-02 18:22:13 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql: tracking issue for advanced planning infrastructure | investigation performance | Jotting down some action items imagined with @RaduBerinde and @andreimatei in response to Alberto's training.
**Main idea: start exercising newly found knowledge / ideas using a tool external to CockroachDB, designed/implemented for learning, but that can be later integrated in CockroachDB.**
- **Action items towards an external tool:**
1. represent logical plans as an IR suitable for experimentation:
1. define a tree language to represent logical plans - this needs to expose the notion of equivalence classes, different algorithms for a single logical operator, and respond to some ancillary semantic concerns (see discussion w/ andrei) [S] @andreimatei, @knz
2. implement an easy-for-the-human syntax to input arbitrary logical plans [S] @knz
3. complete the design & implementation of tree pattern matching and substitution operators [M] @knz @andreimatei (but help is needed)
2. use IR generator to generate some Go code, then make an isolated, out-of-`sql` package to start some preliminary research
1. create data structures & infrastructure to expand logical plans into multiple candidate plans [M]
2. start experimenting?
- how far can we get with experimenting/learning without accurate costing? [Alberto?]
3. iterate to outline/define the external interface of a query optimizer. What are its parameters? [Alberto?]
- can the code that uses it also specify desired ordering?
- can the user code specify whether the plan may contain sorts?
- can the user code express preference for some plans that have some particular properties?
4. iterate to outline/define how is costing resolved? [Alberto?]
- **Work on integration with CockroachDB**:
3. define an interface / architecture to integrate a query optimizer in CockroachDB
1. answer the question: "which interfaces can we use to integrate this in the existing code base?"
3. concurrently/independently with 1+2, investigate how to make logical planning use *sets* of columns instead of ordered lists, so that relational algebraic substitutions become immune to order mismatch
1. define the appropriate way to encode this during logical planning, perhaps start updating the current code to think in that direction [S]
2. design + implement an algorithm that finalizes the decision of a column list in the physical plan at each level of the plan (in which order the columns as placed side-by-side) [M] | True | sql: tracking issue for advanced planning infrastructure - Jotting down some action items imagined with @RaduBerinde and @andreimatei in response to Alberto's training.
**Main idea: start exercising newly found knowledge / ideas using a tool external to CockroachDB, designed/implemented for learning, but that can be later integrated in CockroachDB.**
- **Action items towards an external tool:**
1. represent logical plans as an IR suitable for experimentation:
1. define a tree language to represent logical plans - this needs to expose the notion of equivalence classes, different algorithms for a single logical operator, and respond to some ancillary semantic concerns (see discussion w/ andrei) [S] @andreimatei, @knz
2. implement an easy-for-the-human syntax to input arbitrary logical plans [S] @knz
3. complete the design & implementation of tree pattern matching and substitution operators [M] @knz @andreimatei (but help is needed)
2. use IR generator to generate some Go code, then make an isolated, out-of-`sql` package to start some preliminary research
1. create data structures & infrastructure to expand logical plans into multiple candidate plans [M]
2. start experimenting?
- how far can we get with experimenting/learning without accurate costing? [Alberto?]
3. iterate to outline/define the external interface of a query optimizer. What are its parameters? [Alberto?]
- can the code that uses it also specify desired ordering?
- can the user code specify whether the plan may contain sorts?
- can the user code express preference for some plans that have some particular properties?
4. iterate to outline/define how is costing resolved? [Alberto?]
- **Work on integration with CockroachDB**:
3. define an interface / architecture to integrate a query optimizer in CockroachDB
1. answer the question: "which interfaces can we use to integrate this in the existing code base?"
3. concurrently/independently with 1+2, investigate how to make logical planning use *sets* of columns instead of ordered lists, so that relational algebraic substitutions become immune to order mismatch
1. define the appropriate way to encode this during logical planning, perhaps start updating the current code to think in that direction [S]
2. design + implement an algorithm that finalizes the decision of a column list in the physical plan at each level of the plan (in which order the columns as placed side-by-side) [M] | non_test | sql tracking issue for advanced planning infrastructure jotting down some action items imagined with raduberinde and andreimatei in response to alberto s training main idea start exercising newly found knowledge ideas using a tool external to cockroachdb designed implemented for learning but that can be later integrated in cockroachdb action items towards an external tool represent logical plans as an ir suitable for experimentation define a tree language to represent logical plans this needs to expose the notion of equivalence classes different algorithms for a single logical operator and respond to some ancillary semantic concerns see discussion w andrei andreimatei knz implement an easy for the human syntax to input arbitrary logical plans knz complete the design implementation of tree pattern matching and substitution operators knz andreimatei but help is needed use ir generator to generate some go code then make an isolated out of sql package to start some preliminary research create data structures infrastructure to expand logical plans into multiple candidate plans start experimenting how far can we get with experimenting learning without accurate costing iterate to outline define the external interface of a query optimizer what are its parameters can the code that uses it also specify desired ordering can the user code specify whether the plan may contain sorts can the user code express preference for some plans that have some particular properties iterate to outline define how is costing resolved work on integration with cockroachdb define an interface architecture to integrate a query optimizer in cockroachdb answer the question which interfaces can we use to integrate this in the existing code base concurrently independently with investigate how to make logical planning use sets of columns instead of ordered lists so that relational algebraic substitutions become immune to order mismatch define the appropriate way to encode this during logical planning perhaps start updating the current code to think in that direction design implement an algorithm that finalizes the decision of a column list in the physical plan at each level of the plan in which order the columns as placed side by side | 0 |
141,473 | 11,421,632,254 | IssuesEvent | 2020-02-03 12:37:29 | aconno/Sensorics | https://api.github.com/repos/aconno/Sensorics | opened | Fix ignored unit tests | test | ## Description
Some of the unit tests were failing so we ignored them. Those need to be fixed. | 1.0 | Fix ignored unit tests - ## Description
Some of the unit tests were failing so we ignored them. Those need to be fixed. | test | fix ignored unit tests description some of the unit tests were failing so we ignored them those need to be fixed | 1 |
149,491 | 11,902,893,244 | IssuesEvent | 2020-03-30 14:35:34 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/uptime/settings·ts - Uptime app with generated data uptime settings page changing index pattern setting is reflected elsewhere in UI | Team:uptime [zube]: In Progress blocker failed-test v7.7.0 v8.0.0 | A test failed on a tracked branch
```
{ Error: expected { heartbeatIndices: '' } to sort of equal { heartbeatIndices: 'new*' }
at Assertion.assert (/dev/shm/workspace/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/dev/shm/workspace/kibana/packages/kbn-expect/expect.js:244:8)
at Context.it (test/functional/apps/uptime/settings.ts:68:25)
at process._tickCallback (internal/process/next_tick.js:68:7)
actual: '{\n "heartbeatIndices": ""\n}',
expected: '{\n "heartbeatIndices": "new*"\n}',
showDiff: true }
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3844/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/uptime/settings·ts","test.name":"Uptime app with generated data uptime settings page changing index pattern setting is reflected elsewhere in UI","test.failCount":13}} --> | 1.0 | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/uptime/settings·ts - Uptime app with generated data uptime settings page changing index pattern setting is reflected elsewhere in UI - A test failed on a tracked branch
```
{ Error: expected { heartbeatIndices: '' } to sort of equal { heartbeatIndices: 'new*' }
at Assertion.assert (/dev/shm/workspace/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/dev/shm/workspace/kibana/packages/kbn-expect/expect.js:244:8)
at Context.it (test/functional/apps/uptime/settings.ts:68:25)
at process._tickCallback (internal/process/next_tick.js:68:7)
actual: '{\n "heartbeatIndices": ""\n}',
expected: '{\n "heartbeatIndices": "new*"\n}',
showDiff: true }
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3844/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/uptime/settings·ts","test.name":"Uptime app with generated data uptime settings page changing index pattern setting is reflected elsewhere in UI","test.failCount":13}} --> | test | failing test chrome x pack ui functional tests x pack test functional apps uptime settings·ts uptime app with generated data uptime settings page changing index pattern setting is reflected elsewhere in ui a test failed on a tracked branch error expected heartbeatindices to sort of equal heartbeatindices new at assertion assert dev shm workspace kibana packages kbn expect expect js at assertion eql dev shm workspace kibana packages kbn expect expect js at context it test functional apps uptime settings ts at process tickcallback internal process next tick js actual n heartbeatindices n expected n heartbeatindices new n showdiff true first failure | 1 |
92,784 | 8,378,178,183 | IssuesEvent | 2018-10-06 11:18:20 | curl/curl | https://api.github.com/repos/curl/curl | opened | test 2100 - DoH from command line - is flaky on travis | name lookup tests | See [this build](https://travis-ci.org/curl/curl/builds/437809638?utm_source=github_status&utm_medium=notification) as an example set where two of the Travis-CI builds fail test 2100 (event-based).
The tests *usually* build fine if we retrigger the tests.
I have not managed to trigger this test failure in my own development environments!
### curl/libcurl version
current git master, and DoH support has not been in a release yet
### operating system
Linux, at least
### Test fail log.
I copied the details from the first of the build failures linked to above and it can be view here: [fail-2100.txt](https://github.com/curl/curl/files/2452983/fail-2100.txt)
### Theory
Looking at the log, it might be a case where the requests are sent over the wire in reversed order... | 1.0 | test 2100 - DoH from command line - is flaky on travis - See [this build](https://travis-ci.org/curl/curl/builds/437809638?utm_source=github_status&utm_medium=notification) as an example set where two of the Travis-CI builds fail test 2100 (event-based).
The tests *usually* build fine if we retrigger the tests.
I have not managed to trigger this test failure in my own development environments!
### curl/libcurl version
current git master, and DoH support has not been in a release yet
### operating system
Linux, at least
### Test fail log.
I copied the details from the first of the build failures linked to above and it can be view here: [fail-2100.txt](https://github.com/curl/curl/files/2452983/fail-2100.txt)
### Theory
Looking at the log, it might be a case where the requests are sent over the wire in reversed order... | test | test doh from command line is flaky on travis see as an example set where two of the travis ci builds fail test event based the tests usually build fine if we retrigger the tests i have not managed to trigger this test failure in my own development environments curl libcurl version current git master and doh support has not been in a release yet operating system linux at least test fail log i copied the details from the first of the build failures linked to above and it can be view here theory looking at the log it might be a case where the requests are sent over the wire in reversed order | 1 |
792,759 | 27,973,810,280 | IssuesEvent | 2023-03-25 10:32:29 | SkriptLang/Skript | https://api.github.com/repos/SkriptLang/Skript | reopened | Skript slow to parse "play" syntaxes | bug enhancement priority: medium | ### Skript/Server Version
```
[11:44:00 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases
[11:44:00 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript
[11:44:00 INFO]: [Skript] Server Version: git-Paper-408 (MC: 1.17.1)
[11:44:00 INFO]: [Skript] Skript Version: 2.6.1
[11:44:00 INFO]: [Skript] Installed Skript Addons:
[11:44:00 INFO]: [Skript] - skript-placeholders v1.5.2 (https://github.com/APickledWalrus/skript-placeholders)
[11:44:00 INFO]: [Skript] - skript-db v1.3.3
[11:44:00 INFO]: [Skript] - SkBee v1.14.1 (https://github.com/ShaneBeee/SkBee)
[11:44:00 INFO]: [Skript] - Skript-Packet v2.0.1 (www.github.com/Anarchick/skript-packet)
[11:44:00 INFO]: [Skript] - ItemsAdder v2.5.5 (devs.beer)
[11:44:00 INFO]: [Skript] Installed dependencies:
[11:44:00 INFO]: [Skript] - Vault v1.7.3-b131
[11:44:00 INFO]: [Skript] - WorldGuard v7.0.6+1cff55f
```
### Bug Description
When parsing a skript containing this code skript just crashes. The server still responds.
```
function magicMan(p: player, msg: text):
if {magicman} is not true:
set {magicman} to true
send "&d%{_p}%&8: &f%{_msg}%" to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
set {_answer} to a random integer between 1 and 3
if {_answer} is 1:
send "&dMagicMan&8: &fyes" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1.7 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 2:
send "&dMagicMan&8: &fno" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 0.4 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 3:
send "&dMagicMan&8: &fmaybe" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1 to to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
set {magicman} to false
else:
send "&7..." to {_p}
```
When removing the following part everything works fine:
```
wait 2 seconds
set {_answer} to a random integer between 1 and 3
if {_answer} is 1:
send "&dMagicMan&8: &fyes" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1.7 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 2:
send "&dMagicMan&8: &fno" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 0.4 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 3:
send "&dMagicMan&8: &fmaybe" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1 to to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
```
Even the skUnity fails to parse this code: "Something went wrong trying to parse your script! Please refresh and try again"
### Expected Behavior
Parse the code or at least return an error when something is wrong
### Steps to Reproduce
Parse this code, you don't need to run it
### Errors or Screenshots
https://pastebin.com/dpKkWm9d
### Other
_No response_
### Agreement
- [X] I have read the guidelines above and confirm I am following them with this report. | 1.0 | Skript slow to parse "play" syntaxes - ### Skript/Server Version
```
[11:44:00 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases
[11:44:00 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript
[11:44:00 INFO]: [Skript] Server Version: git-Paper-408 (MC: 1.17.1)
[11:44:00 INFO]: [Skript] Skript Version: 2.6.1
[11:44:00 INFO]: [Skript] Installed Skript Addons:
[11:44:00 INFO]: [Skript] - skript-placeholders v1.5.2 (https://github.com/APickledWalrus/skript-placeholders)
[11:44:00 INFO]: [Skript] - skript-db v1.3.3
[11:44:00 INFO]: [Skript] - SkBee v1.14.1 (https://github.com/ShaneBeee/SkBee)
[11:44:00 INFO]: [Skript] - Skript-Packet v2.0.1 (www.github.com/Anarchick/skript-packet)
[11:44:00 INFO]: [Skript] - ItemsAdder v2.5.5 (devs.beer)
[11:44:00 INFO]: [Skript] Installed dependencies:
[11:44:00 INFO]: [Skript] - Vault v1.7.3-b131
[11:44:00 INFO]: [Skript] - WorldGuard v7.0.6+1cff55f
```
### Bug Description
When parsing a skript containing this code skript just crashes. The server still responds.
```
function magicMan(p: player, msg: text):
if {magicman} is not true:
set {magicman} to true
send "&d%{_p}%&8: &f%{_msg}%" to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
set {_answer} to a random integer between 1 and 3
if {_answer} is 1:
send "&dMagicMan&8: &fyes" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1.7 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 2:
send "&dMagicMan&8: &fno" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 0.4 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 3:
send "&dMagicMan&8: &fmaybe" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1 to to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
set {magicman} to false
else:
send "&7..." to {_p}
```
When removing the following part everything works fine:
```
wait 2 seconds
set {_answer} to a random integer between 1 and 3
if {_answer} is 1:
send "&dMagicMan&8: &fyes" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1.7 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 2:
send "&dMagicMan&8: &fno" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 0.4 to to all players where ["%region at player input%" contains "magicman"]
else if {_answer} is 3:
send "&dMagicMan&8: &fmaybe" to all players where ["%region at player input%" contains "magicman"]
play sound "block.respawn_anchor.charge" with volume 0.8 and pitch 1 to to all players where ["%region at player input%" contains "magicman"]
wait 2 seconds
```
Even the skUnity fails to parse this code: "Something went wrong trying to parse your script! Please refresh and try again"
### Expected Behavior
Parse the code or at least return an error when something is wrong
### Steps to Reproduce
Parse this code, you don't need to run it
### Errors or Screenshots
https://pastebin.com/dpKkWm9d
### Other
_No response_
### Agreement
- [X] I have read the guidelines above and confirm I am following them with this report. | non_test | skript slow to parse play syntaxes skript server version skript s aliases can be found here skript s documentation can be found here server version git paper mc skript version installed skript addons skript placeholders skript db skbee skript packet itemsadder devs beer installed dependencies vault worldguard bug description when parsing a skript containing this code skript just crashes the server still responds function magicman p player msg text if magicman is not true set magicman to true send d p f msg to all players where wait seconds set answer to a random integer between and if answer is send dmagicman fyes to all players where play sound block respawn anchor charge with volume and pitch to to all players where else if answer is send dmagicman fno to all players where play sound block respawn anchor charge with volume and pitch to to all players where else if answer is send dmagicman fmaybe to all players where play sound block respawn anchor charge with volume and pitch to to all players where wait seconds set magicman to false else send to p when removing the following part everything works fine wait seconds set answer to a random integer between and if answer is send dmagicman fyes to all players where play sound block respawn anchor charge with volume and pitch to to all players where else if answer is send dmagicman fno to all players where play sound block respawn anchor charge with volume and pitch to to all players where else if answer is send dmagicman fmaybe to all players where play sound block respawn anchor charge with volume and pitch to to all players where wait seconds even the skunity fails to parse this code something went wrong trying to parse your script please refresh and try again expected behavior parse the code or at least return an error when something is wrong steps to reproduce parse this code you don t need to run it errors or screenshots other no response agreement i have read the guidelines above and confirm i am following them with this report | 0 |
144,925 | 13,130,713,555 | IssuesEvent | 2020-08-06 15:46:57 | goatandsheep/react-native-dotenv | https://api.github.com/repos/goatandsheep/react-native-dotenv | closed | Relation to zetachang/react-native-dotenv - migration guide? | documentation | For a longer-time `react-native-dotenv` user it is very confusing to suddenly have a different repo and no `breaking changes` or `migration` mentions in the readme.
Is this the same/related package? Why was this moved? | 1.0 | Relation to zetachang/react-native-dotenv - migration guide? - For a longer-time `react-native-dotenv` user it is very confusing to suddenly have a different repo and no `breaking changes` or `migration` mentions in the readme.
Is this the same/related package? Why was this moved? | non_test | relation to zetachang react native dotenv migration guide for a longer time react native dotenv user it is very confusing to suddenly have a different repo and no breaking changes or migration mentions in the readme is this the same related package why was this moved | 0 |
79,175 | 9,850,862,484 | IssuesEvent | 2019-06-19 09:09:08 | nextcloud/server | https://api.github.com/repos/nextcloud/server | closed | Auto focus after clicking on share shows drop down list and scrolling issue | 1. to develop backport-request bug design papercut | Sometimes you just would like to open the file inspector to copy an already created shared link or to display file details. By clicking on the file (or even by clicking on "details"), the auto focus on "username or e-mail" fires the drop down list with users and groups to select. This hides your created shared links.
The behaviour is becoming worse if you start scrolling. The drop down has no fix position next to the input textfiled. Instead it starts scrolling as well.
### Steps to reproduce
1. click on a file in list (or on share, or on date, or on "details")
2. file inspector is shown
### Expected behaviour
nothing else
### Actual behaviour
1. auto focus on input filed -> drop down list appears --> is hiding essential information
2. on scrolling the list is also scrolling
### Server configuration
**Operating system**: Debian 10
**Web server:** Apache/2.4.38
**Database:** MariaDB 10.3.14
**PHP version:** PHP 7.3.4
**Nextcloud version:** 16.0.0
**Updated from an older Nextcloud/ownCloud or fresh install:** update
**Where did you install Nextcloud from:** web updater
**Signing status:**
<details>
<summary>Signing status</summary>
No errors have been found.
</details>
**List of activated apps:**
<details>
<summary>App list</summary>
Enabled:
- accessibility: 1.2.0
- activity: 2.9.1
- audioplayer: 2.7.0
- cloud_federation_api: 0.2.0
- comments: 1.6.0
- dav: 1.9.2
- federatedfilesharing: 1.6.0
- files: 1.11.0
- files_pdfviewer: 1.5.0
- files_rightclick: 0.13.0
- files_sharing: 1.8.0
- files_texteditor: 2.8.0
- files_trashbin: 1.6.0
- files_versions: 1.9.0
- files_videoplayer: 1.5.0
- gallery: 18.3.0
- logreader: 2.1.0
- lookup_server_connector: 1.4.0
- nextcloud_announcements: 1.5.0
- notifications: 2.4.1
- oauth2: 1.4.2
- password_policy: 1.6.0
- privacy: 1.0.0
- provisioning_api: 1.6.0
- recommendations: 0.4.0
- serverinfo: 1.6.0
- sharebymail: 1.6.0
- support: 1.0.0
- survey_client: 1.4.0
- systemtags: 1.6.0
- theming: 1.7.0
- twofactor_backupcodes: 1.5.0
- updatenotification: 1.6.0
- viewer: 1.0.0
- workflowengine: 1.6.0
Disabled:
- admin_audit
- encryption
- federation
- files_external
- firstrunwizard
- user_ldap
</details>
**Are you using external storage, if yes which one:** local
**Are you using encryption:** yes
**Are you using an external user-backend, if yes which one:** no
| 1.0 | Auto focus after clicking on share shows drop down list and scrolling issue - Sometimes you just would like to open the file inspector to copy an already created shared link or to display file details. By clicking on the file (or even by clicking on "details"), the auto focus on "username or e-mail" fires the drop down list with users and groups to select. This hides your created shared links.
The behaviour is becoming worse if you start scrolling. The drop down has no fix position next to the input textfiled. Instead it starts scrolling as well.
### Steps to reproduce
1. click on a file in list (or on share, or on date, or on "details")
2. file inspector is shown
### Expected behaviour
nothing else
### Actual behaviour
1. auto focus on input filed -> drop down list appears --> is hiding essential information
2. on scrolling the list is also scrolling
### Server configuration
**Operating system**: Debian 10
**Web server:** Apache/2.4.38
**Database:** MariaDB 10.3.14
**PHP version:** PHP 7.3.4
**Nextcloud version:** 16.0.0
**Updated from an older Nextcloud/ownCloud or fresh install:** update
**Where did you install Nextcloud from:** web updater
**Signing status:**
<details>
<summary>Signing status</summary>
No errors have been found.
</details>
**List of activated apps:**
<details>
<summary>App list</summary>
Enabled:
- accessibility: 1.2.0
- activity: 2.9.1
- audioplayer: 2.7.0
- cloud_federation_api: 0.2.0
- comments: 1.6.0
- dav: 1.9.2
- federatedfilesharing: 1.6.0
- files: 1.11.0
- files_pdfviewer: 1.5.0
- files_rightclick: 0.13.0
- files_sharing: 1.8.0
- files_texteditor: 2.8.0
- files_trashbin: 1.6.0
- files_versions: 1.9.0
- files_videoplayer: 1.5.0
- gallery: 18.3.0
- logreader: 2.1.0
- lookup_server_connector: 1.4.0
- nextcloud_announcements: 1.5.0
- notifications: 2.4.1
- oauth2: 1.4.2
- password_policy: 1.6.0
- privacy: 1.0.0
- provisioning_api: 1.6.0
- recommendations: 0.4.0
- serverinfo: 1.6.0
- sharebymail: 1.6.0
- support: 1.0.0
- survey_client: 1.4.0
- systemtags: 1.6.0
- theming: 1.7.0
- twofactor_backupcodes: 1.5.0
- updatenotification: 1.6.0
- viewer: 1.0.0
- workflowengine: 1.6.0
Disabled:
- admin_audit
- encryption
- federation
- files_external
- firstrunwizard
- user_ldap
</details>
**Are you using external storage, if yes which one:** local
**Are you using encryption:** yes
**Are you using an external user-backend, if yes which one:** no
| non_test | auto focus after clicking on share shows drop down list and scrolling issue sometimes you just would like to open the file inspector to copy an already created shared link or to display file details by clicking on the file or even by clicking on details the auto focus on username or e mail fires the drop down list with users and groups to select this hides your created shared links the behaviour is becoming worse if you start scrolling the drop down has no fix position next to the input textfiled instead it starts scrolling as well steps to reproduce click on a file in list or on share or on date or on details file inspector is shown expected behaviour nothing else actual behaviour auto focus on input filed drop down list appears is hiding essential information on scrolling the list is also scrolling server configuration operating system debian web server apache database mariadb php version php nextcloud version updated from an older nextcloud owncloud or fresh install update where did you install nextcloud from web updater signing status signing status no errors have been found list of activated apps app list enabled accessibility activity audioplayer cloud federation api comments dav federatedfilesharing files files pdfviewer files rightclick files sharing files texteditor files trashbin files versions files videoplayer gallery logreader lookup server connector nextcloud announcements notifications password policy privacy provisioning api recommendations serverinfo sharebymail support survey client systemtags theming twofactor backupcodes updatenotification viewer workflowengine disabled admin audit encryption federation files external firstrunwizard user ldap are you using external storage if yes which one local are you using encryption yes are you using an external user backend if yes which one no | 0 |
250,425 | 27,086,700,229 | IssuesEvent | 2023-02-14 17:30:29 | sharad16j/sharad16j.github.io | https://api.github.com/repos/sharad16j/sharad16j.github.io | opened | CVE-2018-14040 (Medium) detected in bootstrap-3.3.4.min.js | security vulnerability | ## CVE-2018-14040 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.4.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sharad16j/sharad16j.github.io/commit/d8cb57e2f62b9467e3f5334af04eaf318de29b49">d8cb57e2f62b9467e3f5334af04eaf318de29b49</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-14040 (Medium) detected in bootstrap-3.3.4.min.js - ## CVE-2018-14040 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.4.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sharad16j/sharad16j.github.io/commit/d8cb57e2f62b9467e3f5334af04eaf318de29b49">d8cb57e2f62b9467e3f5334af04eaf318de29b49</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file index html path to vulnerable library index html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with mend | 0 |
33,842 | 4,863,481,622 | IssuesEvent | 2016-11-14 15:34:26 | golang/go | https://api.github.com/repos/golang/go | closed | runtime: TestGdbBacktrace fails on netbsd/amd64 | NeedsInvestigation OS-NetBSD Testing | git rev-parse HEAD: 0b6e5e3d733c1da53244753b42940eddb7401c6c
all.bash on netbsd 7.0 (amd64) fails with the following:
```
--- FAIL: TestGdbBacktrace (1.25s)
runtime-gdb_test.go:46: gdb version 7.7
runtime-gdb_test.go:250: could not find '#0.*main\.eee' in backtrace
runtime-gdb_test.go:251: gdb output:
Breakpoint 1 at 0x4010d0: file /tmp/go-build516897324/main.go, line 17.
Program received signal SIGTRAP, Trace/breakpoint trap.
runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
57 MOVL AX, ret+32(FP)
#0 runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
#1 0x000000000041e442 in runtime.semasleep (ns=-1, ~r1=0) at /go/src/runtime/os_netbsd.go:124
#2 0x00000000004092f3 in runtime.notesleep (n=0xc420026518) at /go/src/runtime/lock_sema.go:166
#3 0x00000000004258b3 in runtime.stopm () at /go/src/runtime/proc.go:1576
#4 0x00000000004260c5 in runtime.startlockedm (gp=0xc4200001a0) at /go/src/runtime/proc.go:1753
#5 0x000000000042709a in runtime.schedule () at /go/src/runtime/proc.go:2115
#6 0x0000000000427449 in runtime.park_m (gp=0xc420000820) at /go/src/runtime/proc.go:2165
#7 0x000000000044407b in runtime.mcall () at /go/src/runtime/asm_amd64.s:240
#8 0x000000c420018000 in ?? ()
#9 0x0000000000000000 in ?? ()
Program received signal SIGTRAP, Trace/breakpoint trap.
runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
57 MOVL AX, ret+32(FP)
FAIL
FAIL runtime 41.243s
```
| 1.0 | runtime: TestGdbBacktrace fails on netbsd/amd64 - git rev-parse HEAD: 0b6e5e3d733c1da53244753b42940eddb7401c6c
all.bash on netbsd 7.0 (amd64) fails with the following:
```
--- FAIL: TestGdbBacktrace (1.25s)
runtime-gdb_test.go:46: gdb version 7.7
runtime-gdb_test.go:250: could not find '#0.*main\.eee' in backtrace
runtime-gdb_test.go:251: gdb output:
Breakpoint 1 at 0x4010d0: file /tmp/go-build516897324/main.go, line 17.
Program received signal SIGTRAP, Trace/breakpoint trap.
runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
57 MOVL AX, ret+32(FP)
#0 runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
#1 0x000000000041e442 in runtime.semasleep (ns=-1, ~r1=0) at /go/src/runtime/os_netbsd.go:124
#2 0x00000000004092f3 in runtime.notesleep (n=0xc420026518) at /go/src/runtime/lock_sema.go:166
#3 0x00000000004258b3 in runtime.stopm () at /go/src/runtime/proc.go:1576
#4 0x00000000004260c5 in runtime.startlockedm (gp=0xc4200001a0) at /go/src/runtime/proc.go:1753
#5 0x000000000042709a in runtime.schedule () at /go/src/runtime/proc.go:2115
#6 0x0000000000427449 in runtime.park_m (gp=0xc420000820) at /go/src/runtime/proc.go:2165
#7 0x000000000044407b in runtime.mcall () at /go/src/runtime/asm_amd64.s:240
#8 0x000000c420018000 in ?? ()
#9 0x0000000000000000 in ?? ()
Program received signal SIGTRAP, Trace/breakpoint trap.
runtime.lwp_park () at /go/src/runtime/sys_netbsd_amd64.s:57
57 MOVL AX, ret+32(FP)
FAIL
FAIL runtime 41.243s
```
| test | runtime testgdbbacktrace fails on netbsd git rev parse head all bash on netbsd fails with the following fail testgdbbacktrace runtime gdb test go gdb version runtime gdb test go could not find main eee in backtrace runtime gdb test go gdb output breakpoint at file tmp go main go line program received signal sigtrap trace breakpoint trap runtime lwp park at go src runtime sys netbsd s movl ax ret fp runtime lwp park at go src runtime sys netbsd s in runtime semasleep ns at go src runtime os netbsd go in runtime notesleep n at go src runtime lock sema go in runtime stopm at go src runtime proc go in runtime startlockedm gp at go src runtime proc go in runtime schedule at go src runtime proc go in runtime park m gp at go src runtime proc go in runtime mcall at go src runtime asm s in in program received signal sigtrap trace breakpoint trap runtime lwp park at go src runtime sys netbsd s movl ax ret fp fail fail runtime | 1 |
272,507 | 23,677,613,296 | IssuesEvent | 2022-08-28 10:23:04 | lukaszkukla/safely | https://api.github.com/repos/lukaszkukla/safely | closed | TEST: CRUD categories, risks and statuses on Heroku - admin user | test | Test CRUD categories for admin user
- only admin user is allowed to interact with categories, risks and statuses, links are visible and can be access through browsers' address bar

| 1.0 | TEST: CRUD categories, risks and statuses on Heroku - admin user - Test CRUD categories for admin user
- only admin user is allowed to interact with categories, risks and statuses, links are visible and can be access through browsers' address bar

| test | test crud categories risks and statuses on heroku admin user test crud categories for admin user only admin user is allowed to interact with categories risks and statuses links are visible and can be access through browsers address bar | 1 |
503,560 | 14,594,383,320 | IssuesEvent | 2020-12-20 05:21:09 | myConsciousness/entity-validator | https://api.github.com/repos/myConsciousness/entity-validator | closed | FieldHelperにジェネリクスの型を返却する機能の追加 | Priority: high Type: new feature | # Add New Feature
## 1. Feature details
EnvaliFieldHelper にフィールドのジェネリクス型を返却する機能を追加する。
## 2. Why it is necessary
既存機能の拡張に伴う修正。
## 3. How to implement
リフレクションでジェネリクス型を返却する処理を追加する。
## 4. References
| 1.0 | FieldHelperにジェネリクスの型を返却する機能の追加 - # Add New Feature
## 1. Feature details
EnvaliFieldHelper にフィールドのジェネリクス型を返却する機能を追加する。
## 2. Why it is necessary
既存機能の拡張に伴う修正。
## 3. How to implement
リフレクションでジェネリクス型を返却する処理を追加する。
## 4. References
| non_test | fieldhelperにジェネリクスの型を返却する機能の追加 add new feature feature details envalifieldhelper にフィールドのジェネリクス型を返却する機能を追加する。 why it is necessary 既存機能の拡張に伴う修正。 how to implement リフレクションでジェネリクス型を返却する処理を追加する。 references | 0 |
176,881 | 13,657,181,941 | IssuesEvent | 2020-09-28 05:05:06 | tarantool/tarantool | https://api.github.com/repos/tarantool/tarantool | opened | test: flaky replication/wal_rw_stress.test.lua test | flaky test qa | Tarantool version:
Tarantool 2.6.0-115-ge19b392f0
Target: Darwin-x86_64-RelWithDebInfo
Build options: cmake . -DCMAKE_INSTALL_PREFIX=/Users/tntmac04.tarantool.i/builds/PVWCJs8U/0/tarantool/tarantool/static-build/tarantool-prefix -DENABLE_BACKTRACE=TRUE
Compiler: /Library/Developer/CommandLineTools/usr/bin/cc /Library/Developer/CommandLineTools/usr/bin/c++
C_FLAGS: -Wno-unknown-pragmas -fexceptions -funwind-tables -fno-omit-frame-pointer -fno-stack-protector -fno-common -msse2 -std=c11 -Wall -Wextra -Wno-strict-aliasing -Wno-char-subscripts -Wno-gnu-alignof-expression -Werror
CXX_FLAGS: -Wno-unknown-pragmas -fexceptions -funwind-tables -fno-omit-frame-pointer -fno-stack-protector -fno-common -msse2 -std=c++11 -Wall -Wextra -Wno-strict-aliasing -Wno-char-subscripts -Wno-invalid-offsetof -Wno-gnu-alignof-expression -Werror
OS version:
Linux (CentOS 7)
Bug description:
https://gitlab.com/tarantool/tarantool/-/jobs/759812885#L6023
[artifacts.zip](https://github.com/tarantool/tarantool/files/5290000/artifacts.zip)
Not possible to use results file checksum due to different data in error message:
```
[048] --- replication/wal_rw_stress.result Wed Aug 26 06:06:16 2020
[048] +++ replication/wal_rw_stress.reject Mon Sep 28 03:16:57 2020
[048] @@ -77,7 +77,40 @@
[048] r.downstream.status ~= 'stopped') \
[048] end) or box.info
[048] ---
[048] -- true
[048] +- version: 2.6.0-115-ge19b392f0
[048] + id: 2
[048] + ro: false
[048] + uuid: 63104370-9fc6-4162-b5ca-b2ae6be2e179
[048] + package: Tarantool
[048] + cluster:
[048] + uuid: 55043ca8-d79b-4a89-8551-db0c0059a80e
[048] + listen: unix/:/tmp/tnt/048_replication/replica.socket-iproto
[048] + replication_anon:
[048] + count: 0
[048] + replication:
[048] + 1:
[048] + id: 1
[048] + uuid: 0eeabfcd-723f-4e1d-b0c2-bb820c3a6657
[048] + lsn: 10636
[048] + upstream:
[048] + status: follow
[048] + idle: 0.20655800029635
[048] + peer: unix/:/tmp/tnt/048_replication/master.socket-iproto
[048] + lag: 9.918212890625e-05
[048] + 2:
[048] + id: 2
[048] + uuid: 63104370-9fc6-4162-b5ca-b2ae6be2e179
[048] + lsn: 0
[048] + signature: 10636
[048] + status: running
[048] + vclock: {1: 10636}
[048] + uptime: 61
[048] + lsn: 0
[048] + sql: []
[048] + gc: []
[048] + vinyl: []
[048] + memory: []
[048] + pid: 33722
[048] ...
[048] test_run:cmd("switch default")
[048] ---
[048]
```
Steps to reproduce:
Optional (but very desirable):
* coredump
* backtrace
* netstat
| 1.0 | test: flaky replication/wal_rw_stress.test.lua test - Tarantool version:
Tarantool 2.6.0-115-ge19b392f0
Target: Darwin-x86_64-RelWithDebInfo
Build options: cmake . -DCMAKE_INSTALL_PREFIX=/Users/tntmac04.tarantool.i/builds/PVWCJs8U/0/tarantool/tarantool/static-build/tarantool-prefix -DENABLE_BACKTRACE=TRUE
Compiler: /Library/Developer/CommandLineTools/usr/bin/cc /Library/Developer/CommandLineTools/usr/bin/c++
C_FLAGS: -Wno-unknown-pragmas -fexceptions -funwind-tables -fno-omit-frame-pointer -fno-stack-protector -fno-common -msse2 -std=c11 -Wall -Wextra -Wno-strict-aliasing -Wno-char-subscripts -Wno-gnu-alignof-expression -Werror
CXX_FLAGS: -Wno-unknown-pragmas -fexceptions -funwind-tables -fno-omit-frame-pointer -fno-stack-protector -fno-common -msse2 -std=c++11 -Wall -Wextra -Wno-strict-aliasing -Wno-char-subscripts -Wno-invalid-offsetof -Wno-gnu-alignof-expression -Werror
OS version:
Linux (CentOS 7)
Bug description:
https://gitlab.com/tarantool/tarantool/-/jobs/759812885#L6023
[artifacts.zip](https://github.com/tarantool/tarantool/files/5290000/artifacts.zip)
Not possible to use results file checksum due to different data in error message:
```
[048] --- replication/wal_rw_stress.result Wed Aug 26 06:06:16 2020
[048] +++ replication/wal_rw_stress.reject Mon Sep 28 03:16:57 2020
[048] @@ -77,7 +77,40 @@
[048] r.downstream.status ~= 'stopped') \
[048] end) or box.info
[048] ---
[048] -- true
[048] +- version: 2.6.0-115-ge19b392f0
[048] + id: 2
[048] + ro: false
[048] + uuid: 63104370-9fc6-4162-b5ca-b2ae6be2e179
[048] + package: Tarantool
[048] + cluster:
[048] + uuid: 55043ca8-d79b-4a89-8551-db0c0059a80e
[048] + listen: unix/:/tmp/tnt/048_replication/replica.socket-iproto
[048] + replication_anon:
[048] + count: 0
[048] + replication:
[048] + 1:
[048] + id: 1
[048] + uuid: 0eeabfcd-723f-4e1d-b0c2-bb820c3a6657
[048] + lsn: 10636
[048] + upstream:
[048] + status: follow
[048] + idle: 0.20655800029635
[048] + peer: unix/:/tmp/tnt/048_replication/master.socket-iproto
[048] + lag: 9.918212890625e-05
[048] + 2:
[048] + id: 2
[048] + uuid: 63104370-9fc6-4162-b5ca-b2ae6be2e179
[048] + lsn: 0
[048] + signature: 10636
[048] + status: running
[048] + vclock: {1: 10636}
[048] + uptime: 61
[048] + lsn: 0
[048] + sql: []
[048] + gc: []
[048] + vinyl: []
[048] + memory: []
[048] + pid: 33722
[048] ...
[048] test_run:cmd("switch default")
[048] ---
[048]
```
Steps to reproduce:
Optional (but very desirable):
* coredump
* backtrace
* netstat
| test | test flaky replication wal rw stress test lua test tarantool version tarantool target darwin relwithdebinfo build options cmake dcmake install prefix users tarantool i builds tarantool tarantool static build tarantool prefix denable backtrace true compiler library developer commandlinetools usr bin cc library developer commandlinetools usr bin c c flags wno unknown pragmas fexceptions funwind tables fno omit frame pointer fno stack protector fno common std wall wextra wno strict aliasing wno char subscripts wno gnu alignof expression werror cxx flags wno unknown pragmas fexceptions funwind tables fno omit frame pointer fno stack protector fno common std c wall wextra wno strict aliasing wno char subscripts wno invalid offsetof wno gnu alignof expression werror os version linux centos bug description not possible to use results file checksum due to different data in error message replication wal rw stress result wed aug replication wal rw stress reject mon sep r downstream status stopped end or box info true version id ro false uuid package tarantool cluster uuid listen unix tmp tnt replication replica socket iproto replication anon count replication id uuid lsn upstream status follow idle peer unix tmp tnt replication master socket iproto lag id uuid lsn signature status running vclock uptime lsn sql gc vinyl memory pid test run cmd switch default steps to reproduce optional but very desirable coredump backtrace netstat | 1 |
186,334 | 6,735,325,679 | IssuesEvent | 2017-10-18 21:20:28 | regnauld/netdot-redmine-test | https://api.github.com/repos/regnauld/netdot-redmine-test | closed | Extend usage statistics to other block types besides "Subnet". | Priority 4: Normal Tracker 1: Bug | ---
Author Name: **Carlos Vicente** (Carlos Vicente)
Original Redmine Issue: 8, http://localhost:3000/issues/8
Original Assignee: Carlos Vicente
---
Extend usage statistics to other block types besides "Subnet".
| 1.0 | Extend usage statistics to other block types besides "Subnet". - ---
Author Name: **Carlos Vicente** (Carlos Vicente)
Original Redmine Issue: 8, http://localhost:3000/issues/8
Original Assignee: Carlos Vicente
---
Extend usage statistics to other block types besides "Subnet".
| non_test | extend usage statistics to other block types besides subnet author name carlos vicente carlos vicente original redmine issue original assignee carlos vicente extend usage statistics to other block types besides subnet | 0 |
232,316 | 18,857,669,849 | IssuesEvent | 2021-11-12 08:53:32 | mattermost/mattermost-server | https://api.github.com/repos/mattermost/mattermost-server | closed | Write Webapp E2E with Cypress: "MM-T2074 New email not visible to other users until it has been confirmed" | Difficulty/1:Easy Hacktoberfest Area/E2E Tests Help Wanted Tech/Automation | See our [documentation for Webapp end-to-end testing with Cypress](https://developers.mattermost.com/contribute/webapp/end-to-end-tests/) for reference.
<article>
<h1>MM-T2074 New email not visible to other users until it has been confirmed</h1>
<div>
<div>
<h3>Steps </h3>New email not visible to oher users until it has been confirmed <br />————————————————————————————<br />1. Login as sys admin and ensure that "Require Email Verification is set to True uder Sign Up in the system console<br />2. Login as a test user <br />3. Update your email address to something new not in use by other users on trhe server - Do not verify the new address<br />4. As the test user send a DM to a 2nd user<br />5. Login as the 2nd user and view the profile popover of the test user<h3>Expected</h3>The test user's profile still shows the old email address<hr>
</div>
</div>
</article>
**Test Folder:** ``/cypress/integration/account_settings``
**Test code arrangement:**
```
describe('Account Settings', () => {
it('MM-T2074 New email not visible to other users until it has been confirmed', () => {
// code
});
});
```
Notes:
1. Do not add ``@prod`` label in a spec file
- If you're writing script into a newly created test file, ``@prod`` label should not be included.
- If you're adding script into an existing test file, ``@prod`` label should removed.
2. Use [queries from testing-library](https://testing-library.com/docs/dom-testing-library/api-queries) whenever possible. We share the same philosophy as the [testing-library](https://testing-library.com/) when doing UI automation like "Interact with your app the same way as your users" and so, please follow their guidelines especially when querying an element.
If you're interested, please comment here and come [join our "Contributors" community channel](https://community.mattermost.com/core/channels/tickets) on our daily build server, where you can discuss questions with community members and the Mattermost core team. For technical advice or questions, please [join our "Developers" community channel](https://community.mattermost.com/core/channels/developers).
New contributors please see our [Developer's Guide](https://developers.mattermost.com/contribute/getting-started/).
| 1.0 | Write Webapp E2E with Cypress: "MM-T2074 New email not visible to other users until it has been confirmed" - See our [documentation for Webapp end-to-end testing with Cypress](https://developers.mattermost.com/contribute/webapp/end-to-end-tests/) for reference.
<article>
<h1>MM-T2074 New email not visible to other users until it has been confirmed</h1>
<div>
<div>
<h3>Steps </h3>New email not visible to oher users until it has been confirmed <br />————————————————————————————<br />1. Login as sys admin and ensure that "Require Email Verification is set to True uder Sign Up in the system console<br />2. Login as a test user <br />3. Update your email address to something new not in use by other users on trhe server - Do not verify the new address<br />4. As the test user send a DM to a 2nd user<br />5. Login as the 2nd user and view the profile popover of the test user<h3>Expected</h3>The test user's profile still shows the old email address<hr>
</div>
</div>
</article>
**Test Folder:** ``/cypress/integration/account_settings``
**Test code arrangement:**
```
describe('Account Settings', () => {
it('MM-T2074 New email not visible to other users until it has been confirmed', () => {
// code
});
});
```
Notes:
1. Do not add ``@prod`` label in a spec file
- If you're writing script into a newly created test file, ``@prod`` label should not be included.
- If you're adding script into an existing test file, ``@prod`` label should removed.
2. Use [queries from testing-library](https://testing-library.com/docs/dom-testing-library/api-queries) whenever possible. We share the same philosophy as the [testing-library](https://testing-library.com/) when doing UI automation like "Interact with your app the same way as your users" and so, please follow their guidelines especially when querying an element.
If you're interested, please comment here and come [join our "Contributors" community channel](https://community.mattermost.com/core/channels/tickets) on our daily build server, where you can discuss questions with community members and the Mattermost core team. For technical advice or questions, please [join our "Developers" community channel](https://community.mattermost.com/core/channels/developers).
New contributors please see our [Developer's Guide](https://developers.mattermost.com/contribute/getting-started/).
| test | write webapp with cypress mm new email not visible to other users until it has been confirmed see our for reference mm new email not visible to other users until it has been confirmed steps new email not visible to oher users until it has been confirmed mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash mdash login as sys admin and ensure that quot require email verification is set to true uder sign up in the system console login as a test user update your email address to something new not in use by other users on trhe server do not verify the new address as the test user send a dm to a user login as the user and view the profile popover of the test user expected the test user s profile still shows the old email address test folder cypress integration account settings test code arrangement describe account settings it mm new email not visible to other users until it has been confirmed code notes do not add prod label in a spec file if you re writing script into a newly created test file prod label should not be included if you re adding script into an existing test file prod label should removed use whenever possible we share the same philosophy as the when doing ui automation like interact with your app the same way as your users and so please follow their guidelines especially when querying an element if you re interested please comment here and come on our daily build server where you can discuss questions with community members and the mattermost core team for technical advice or questions please new contributors please see our | 1 |
46,019 | 9,856,722,856 | IssuesEvent | 2019-06-19 23:17:01 | BrightSpots/rcv | https://api.github.com/repos/BrightSpots/rcv | opened | Determine if TieBreaker pop-up needs parent window | code cleanup enhancement | Comment in `GuiApplication`:
"// cache main window so we can parent file choosers to it"
Probably more accurately make this "parent pop-up windows" in general
See if any other window parenting needs to be done (e.g. for Tiebreaker, alerts, etc.), like this:
https://github.com/BrightSpots/rcv/pull/172/files
| 1.0 | Determine if TieBreaker pop-up needs parent window - Comment in `GuiApplication`:
"// cache main window so we can parent file choosers to it"
Probably more accurately make this "parent pop-up windows" in general
See if any other window parenting needs to be done (e.g. for Tiebreaker, alerts, etc.), like this:
https://github.com/BrightSpots/rcv/pull/172/files
| non_test | determine if tiebreaker pop up needs parent window comment in guiapplication cache main window so we can parent file choosers to it probably more accurately make this parent pop up windows in general see if any other window parenting needs to be done e g for tiebreaker alerts etc like this | 0 |
133,415 | 10,820,921,060 | IssuesEvent | 2019-11-08 17:25:34 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Editor crash when changing nodes in scene panel | bug needs testing topic:editor | **Godot version:**
3.1 8698876
**OS/device including version:**
Ubuntu 18.04.1 GTX 970
**Issue description:**
When I changed selection of node via keyboard, then editor crash and this error appear:
```
[1] /lib/x86_64-linux-gnu/libc.so.6(+0x3ef20) [0x7fa734c92f20] (??:0)
[2] VectorWriteProxy<Variant>::operator[](int) (/home/rafal/Pulpit/godot/./core/vector.h:49 (discriminator 7))
[3] Array::operator[](int) (/home/rafal/Pulpit/godot/core/array.cpp:76)
[4] Control::_edit_set_state(Dictionary const&) (/home/rafal/Pulpit/godot/scene/gui/control.cpp:77)
[5] CanvasItemEditor::_restore_canvas_item_state(List<CanvasItem*, DefaultAllocator>, bool) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:805)
[6] CanvasItemEditor::_gui_input_move(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:1891 (discriminator 3))
[7] CanvasItemEditor::_gui_input_viewport(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:2209)
[8] MethodBind1<Ref<InputEvent> const&>::call(Object*, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/./core/method_bind.gen.inc:729 (discriminator 12))
[9] Object::call(StringName const&, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/core/object.cpp:945 (discriminator 1))
[10] Object::emit_signal(StringName const&, Variant const**, int) (/home/rafal/Pulpit/godot/core/object.cpp:1231 (discriminator 1))
[11] Object::emit_signal(StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/core/object.cpp:1288)
[12] Viewport::_gui_input_event(Ref<InputEvent>) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:2269 (discriminator 2))
[13] Viewport::input(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:2663 (discriminator 2))
[14] Viewport::_vp_input(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:1319)
[15] MethodBind1<Ref<InputEvent> const&>::call(Object*, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/./core/method_bind.gen.inc:729 (discriminator 12))
[16] Object::call(StringName const&, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/core/object.cpp:945 (discriminator 1))
[17] Object::call(StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/core/object.cpp:870)
[18] SceneTree::call_group_flags(unsigned int, StringName const&, StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/scene/main/scene_tree.cpp:262)
[19] SceneTree::input_event(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/scene_tree.cpp:418 (discriminator 6))
[20] InputDefault::_parse_input_event_impl(Ref<InputEvent> const&, bool) (/home/rafal/Pulpit/godot/main/input_default.cpp:414)
[21] InputDefault::parse_input_event(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/main/input_default.cpp:260)
[22] OS_X11::handle_key_event(XKeyEvent*, bool) (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:1778 (discriminator 3))
[23] OS_X11::process_xevents() (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:2258)
[24] OS_X11::run() (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:2973)
[25] /usr/bin/godot(main+0xdc) [0x1156893] (/home/rafal/Pulpit/godot/platform/x11/godot_x11.cpp:56)
[26] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xe7) [0x7fa734c75b97] (??:0)
[27] /usr/bin/godot(_start+0x2a) [0x11566fa] (??:?)
```
**Minimal reproduction project:**
TPS Demo
| 1.0 | Editor crash when changing nodes in scene panel - **Godot version:**
3.1 8698876
**OS/device including version:**
Ubuntu 18.04.1 GTX 970
**Issue description:**
When I changed selection of node via keyboard, then editor crash and this error appear:
```
[1] /lib/x86_64-linux-gnu/libc.so.6(+0x3ef20) [0x7fa734c92f20] (??:0)
[2] VectorWriteProxy<Variant>::operator[](int) (/home/rafal/Pulpit/godot/./core/vector.h:49 (discriminator 7))
[3] Array::operator[](int) (/home/rafal/Pulpit/godot/core/array.cpp:76)
[4] Control::_edit_set_state(Dictionary const&) (/home/rafal/Pulpit/godot/scene/gui/control.cpp:77)
[5] CanvasItemEditor::_restore_canvas_item_state(List<CanvasItem*, DefaultAllocator>, bool) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:805)
[6] CanvasItemEditor::_gui_input_move(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:1891 (discriminator 3))
[7] CanvasItemEditor::_gui_input_viewport(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/editor/plugins/canvas_item_editor_plugin.cpp:2209)
[8] MethodBind1<Ref<InputEvent> const&>::call(Object*, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/./core/method_bind.gen.inc:729 (discriminator 12))
[9] Object::call(StringName const&, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/core/object.cpp:945 (discriminator 1))
[10] Object::emit_signal(StringName const&, Variant const**, int) (/home/rafal/Pulpit/godot/core/object.cpp:1231 (discriminator 1))
[11] Object::emit_signal(StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/core/object.cpp:1288)
[12] Viewport::_gui_input_event(Ref<InputEvent>) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:2269 (discriminator 2))
[13] Viewport::input(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:2663 (discriminator 2))
[14] Viewport::_vp_input(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/viewport.cpp:1319)
[15] MethodBind1<Ref<InputEvent> const&>::call(Object*, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/./core/method_bind.gen.inc:729 (discriminator 12))
[16] Object::call(StringName const&, Variant const**, int, Variant::CallError&) (/home/rafal/Pulpit/godot/core/object.cpp:945 (discriminator 1))
[17] Object::call(StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/core/object.cpp:870)
[18] SceneTree::call_group_flags(unsigned int, StringName const&, StringName const&, Variant const&, Variant const&, Variant const&, Variant const&, Variant const&) (/home/rafal/Pulpit/godot/scene/main/scene_tree.cpp:262)
[19] SceneTree::input_event(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/scene/main/scene_tree.cpp:418 (discriminator 6))
[20] InputDefault::_parse_input_event_impl(Ref<InputEvent> const&, bool) (/home/rafal/Pulpit/godot/main/input_default.cpp:414)
[21] InputDefault::parse_input_event(Ref<InputEvent> const&) (/home/rafal/Pulpit/godot/main/input_default.cpp:260)
[22] OS_X11::handle_key_event(XKeyEvent*, bool) (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:1778 (discriminator 3))
[23] OS_X11::process_xevents() (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:2258)
[24] OS_X11::run() (/home/rafal/Pulpit/godot/platform/x11/os_x11.cpp:2973)
[25] /usr/bin/godot(main+0xdc) [0x1156893] (/home/rafal/Pulpit/godot/platform/x11/godot_x11.cpp:56)
[26] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xe7) [0x7fa734c75b97] (??:0)
[27] /usr/bin/godot(_start+0x2a) [0x11566fa] (??:?)
```
**Minimal reproduction project:**
TPS Demo
| test | editor crash when changing nodes in scene panel godot version os device including version ubuntu gtx issue description when i changed selection of node via keyboard then editor crash and this error appear lib linux gnu libc so vectorwriteproxy operator int home rafal pulpit godot core vector h discriminator array operator int home rafal pulpit godot core array cpp control edit set state dictionary const home rafal pulpit godot scene gui control cpp canvasitemeditor restore canvas item state list bool home rafal pulpit godot editor plugins canvas item editor plugin cpp canvasitemeditor gui input move ref const home rafal pulpit godot editor plugins canvas item editor plugin cpp discriminator canvasitemeditor gui input viewport ref const home rafal pulpit godot editor plugins canvas item editor plugin cpp const call object variant const int variant callerror home rafal pulpit godot core method bind gen inc discriminator object call stringname const variant const int variant callerror home rafal pulpit godot core object cpp discriminator object emit signal stringname const variant const int home rafal pulpit godot core object cpp discriminator object emit signal stringname const variant const variant const variant const variant const variant const home rafal pulpit godot core object cpp viewport gui input event ref home rafal pulpit godot scene main viewport cpp discriminator viewport input ref const home rafal pulpit godot scene main viewport cpp discriminator viewport vp input ref const home rafal pulpit godot scene main viewport cpp const call object variant const int variant callerror home rafal pulpit godot core method bind gen inc discriminator object call stringname const variant const int variant callerror home rafal pulpit godot core object cpp discriminator object call stringname const variant const variant const variant const variant const variant const home rafal pulpit godot core object cpp scenetree call group flags unsigned int stringname const stringname const variant const variant const variant const variant const variant const home rafal pulpit godot scene main scene tree cpp scenetree input event ref const home rafal pulpit godot scene main scene tree cpp discriminator inputdefault parse input event impl ref const bool home rafal pulpit godot main input default cpp inputdefault parse input event ref const home rafal pulpit godot main input default cpp os handle key event xkeyevent bool home rafal pulpit godot platform os cpp discriminator os process xevents home rafal pulpit godot platform os cpp os run home rafal pulpit godot platform os cpp usr bin godot main home rafal pulpit godot platform godot cpp lib linux gnu libc so libc start main usr bin godot start minimal reproduction project tps demo | 1 |
535,815 | 15,699,263,166 | IssuesEvent | 2021-03-26 08:13:53 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Filter not working correctly on certificates section | Priority/High bug console ui | **Describe the Issue:**
Filter option did not work correctly when searching the content with space.
**How To Reproduce:**
1. Go to certificates section in the manage tab
2. Search the content with space
Filter not working correctly

**Expected behavior :**
Based on the contain within a space result will return.
like

**Device Information :**
- Device: PC
- OS: Ubuntu
- Browser + Version : Firefox 85.0
| 1.0 | Filter not working correctly on certificates section - **Describe the Issue:**
Filter option did not work correctly when searching the content with space.
**How To Reproduce:**
1. Go to certificates section in the manage tab
2. Search the content with space
Filter not working correctly

**Expected behavior :**
Based on the contain within a space result will return.
like

**Device Information :**
- Device: PC
- OS: Ubuntu
- Browser + Version : Firefox 85.0
| non_test | filter not working correctly on certificates section describe the issue filter option did not work correctly when searching the content with space how to reproduce go to certificates section in the manage tab search the content with space filter not working correctly expected behavior based on the contain within a space result will return like device information device pc os ubuntu browser version firefox | 0 |
214,454 | 16,590,551,117 | IssuesEvent | 2021-06-01 07:08:56 | ransome1/sleek | https://api.github.com/repos/ransome1/sleek | closed | Call for testing: Cluster based loading of the todo list | call for testing help wanted | **Call for testing**
As the rendering of todo entries is the most resource hungry task in sleek, users with large todo.txt files might be having performance issues, particulary on older hardware. This pre-release introduces a cluster based rendering, where not all items will be rendered at once. Instead the todos are splitted into clusters which will be loaded one at a time once the windows scroll bar has hit the bottom.
@9z0, @jason-mehmel, you guys are handling these huge todo files, right? It would be really lovely to hear about your experiences with this pre-release, please feel free to share your feedback. | 1.0 | Call for testing: Cluster based loading of the todo list - **Call for testing**
As the rendering of todo entries is the most resource hungry task in sleek, users with large todo.txt files might be having performance issues, particulary on older hardware. This pre-release introduces a cluster based rendering, where not all items will be rendered at once. Instead the todos are splitted into clusters which will be loaded one at a time once the windows scroll bar has hit the bottom.
@9z0, @jason-mehmel, you guys are handling these huge todo files, right? It would be really lovely to hear about your experiences with this pre-release, please feel free to share your feedback. | test | call for testing cluster based loading of the todo list call for testing as the rendering of todo entries is the most resource hungry task in sleek users with large todo txt files might be having performance issues particulary on older hardware this pre release introduces a cluster based rendering where not all items will be rendered at once instead the todos are splitted into clusters which will be loaded one at a time once the windows scroll bar has hit the bottom jason mehmel you guys are handling these huge todo files right it would be really lovely to hear about your experiences with this pre release please feel free to share your feedback | 1 |
290,418 | 25,065,538,712 | IssuesEvent | 2022-11-07 07:58:52 | mozilla-mobile/focus-android | https://api.github.com/repos/mozilla-mobile/focus-android | closed | UI test failure - <EraseBrowsingDataTest.trashButtonTest> | eng:ui-test eng:intermittent-test eng:disabled-test | SnackBar should will be deleted after we integrate https://github.com/mozilla-mobile/focus-android/issues/7506 . Please see the comment https://github.com/mozilla-mobile/focus-android/issues/7506#issuecomment-1252232982 . This test should be redone.
| 3.0 | UI test failure - <EraseBrowsingDataTest.trashButtonTest> - SnackBar should will be deleted after we integrate https://github.com/mozilla-mobile/focus-android/issues/7506 . Please see the comment https://github.com/mozilla-mobile/focus-android/issues/7506#issuecomment-1252232982 . This test should be redone.
| test | ui test failure snackbar should will be deleted after we integrate please see the comment this test should be redone | 1 |
258,495 | 22,322,968,601 | IssuesEvent | 2022-06-14 08:12:06 | ntop/ntopng | https://api.github.com/repos/ntop/ntopng | closed | Add Error Code filter in the alerts page | enhancement feature request Ready to Test priority ticket | **What would you like to add or change?**:
Add Error Code filters in the alert page in order to be able to search for an Error code status
| 1.0 | Add Error Code filter in the alerts page - **What would you like to add or change?**:
Add Error Code filters in the alert page in order to be able to search for an Error code status
| test | add error code filter in the alerts page what would you like to add or change add error code filters in the alert page in order to be able to search for an error code status | 1 |
348,088 | 10,438,548,611 | IssuesEvent | 2019-09-18 02:36:44 | AY1920S1-CS2103T-W13-3/main | https://api.github.com/repos/AY1920S1-CS2103T-W13-3/main | opened | As a student who usually has many application windows and browser tabs open at once, the todolist, exam scheduler and note compiler provides me with a singular app. | priority.High type.Story | To manage multiple features which I previously used different applications (and therefore windows) for with regards to organization all under one roof. | 1.0 | As a student who usually has many application windows and browser tabs open at once, the todolist, exam scheduler and note compiler provides me with a singular app. - To manage multiple features which I previously used different applications (and therefore windows) for with regards to organization all under one roof. | non_test | as a student who usually has many application windows and browser tabs open at once the todolist exam scheduler and note compiler provides me with a singular app to manage multiple features which i previously used different applications and therefore windows for with regards to organization all under one roof | 0 |
273,623 | 20,799,779,789 | IssuesEvent | 2022-03-17 12:53:58 | Research-Support-Network/research-support-network.github.io | https://api.github.com/repos/Research-Support-Network/research-support-network.github.io | closed | Summary & Explanations | documentation | Text needed:
- [x] A brief Summary / Statement of Purpose / About section on the landing page
- [x] Other important bits of information (also on the landing page)
- [x] Some info to the entry form
- [x] Some info to the list page
- [x] A README in the repo
- [x] A simple code of conduct
- [x] Some links to other resources
I will write something up today or tomorrow. | 1.0 | Summary & Explanations - Text needed:
- [x] A brief Summary / Statement of Purpose / About section on the landing page
- [x] Other important bits of information (also on the landing page)
- [x] Some info to the entry form
- [x] Some info to the list page
- [x] A README in the repo
- [x] A simple code of conduct
- [x] Some links to other resources
I will write something up today or tomorrow. | non_test | summary explanations text needed a brief summary statement of purpose about section on the landing page other important bits of information also on the landing page some info to the entry form some info to the list page a readme in the repo a simple code of conduct some links to other resources i will write something up today or tomorrow | 0 |
21,506 | 3,900,519,093 | IssuesEvent | 2016-04-18 06:36:34 | VengeanceRising/VengeanceRising | https://api.github.com/repos/VengeanceRising/VengeanceRising | closed | Tutorials: usage of q and e problematic | bug Playtest Priority.medium | You can access old tutorial subtitles even after you progressed in the game, which should not be the case.
Examples:
Inventory tutorial still acessible after returning to diningroom level
First tutorial of mainhall still accesible, even after clearing the library level
Hourglass level tutorial still acessible after returning to storage room | 1.0 | Tutorials: usage of q and e problematic - You can access old tutorial subtitles even after you progressed in the game, which should not be the case.
Examples:
Inventory tutorial still acessible after returning to diningroom level
First tutorial of mainhall still accesible, even after clearing the library level
Hourglass level tutorial still acessible after returning to storage room | test | tutorials usage of q and e problematic you can access old tutorial subtitles even after you progressed in the game which should not be the case examples inventory tutorial still acessible after returning to diningroom level first tutorial of mainhall still accesible even after clearing the library level hourglass level tutorial still acessible after returning to storage room | 1 |
179,852 | 13,907,735,900 | IssuesEvent | 2020-10-20 12:59:37 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Add Custom Credential Type | component:ui_next state:needs_test type:feature | ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
Add feature to add Custom Credential Type
| 1.0 | Add Custom Credential Type - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
Add feature to add Custom Credential Type
| test | add custom credential type issue type feature idea summary add feature to add custom credential type | 1 |
117,663 | 9,955,337,039 | IssuesEvent | 2019-07-05 10:44:39 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | XLSX Export: numeric values | QoL fixed (test & close) small change | > The allocations spreadsheet is downloading numerical files as text. Is it possible to make the numerical fields numeric
via: AS/CS | 1.0 | XLSX Export: numeric values - > The allocations spreadsheet is downloading numerical files as text. Is it possible to make the numerical fields numeric
via: AS/CS | test | xlsx export numeric values the allocations spreadsheet is downloading numerical files as text is it possible to make the numerical fields numeric via as cs | 1 |
316,229 | 27,147,515,088 | IssuesEvent | 2023-02-16 21:17:18 | UWB-Biocomputing/Graphitti | https://api.github.com/repos/UWB-Biocomputing/Graphitti | opened | Problem with binned spike count computation in XML output | cleanup Recorders testing | So, the code below is basically all we save when we run a simulation and produce XML output: the binned spike counts (for 10ms bins). Of course, this only works for neural simulations. We want to go to just saving whatever the recorder is told to save. However, it might be good to preserve this computation temporarily, just as a double-check for regression testing. But it should go. Also, it does the computation in a poor way, because we should never compute the time of an event from the beginning of the simulation; just the interval between events (because simulations can be very long — too long to store in a `double` accurately).
https://github.com/UWB-Biocomputing/Graphitti/blob/b5bd8d3298e67b620f96032cfb5c7f484345ec7e/Simulator/Recorders/XmlRecorder.cpp#L92-L96 | 1.0 | Problem with binned spike count computation in XML output - So, the code below is basically all we save when we run a simulation and produce XML output: the binned spike counts (for 10ms bins). Of course, this only works for neural simulations. We want to go to just saving whatever the recorder is told to save. However, it might be good to preserve this computation temporarily, just as a double-check for regression testing. But it should go. Also, it does the computation in a poor way, because we should never compute the time of an event from the beginning of the simulation; just the interval between events (because simulations can be very long — too long to store in a `double` accurately).
https://github.com/UWB-Biocomputing/Graphitti/blob/b5bd8d3298e67b620f96032cfb5c7f484345ec7e/Simulator/Recorders/XmlRecorder.cpp#L92-L96 | test | problem with binned spike count computation in xml output so the code below is basically all we save when we run a simulation and produce xml output the binned spike counts for bins of course this only works for neural simulations we want to go to just saving whatever the recorder is told to save however it might be good to preserve this computation temporarily just as a double check for regression testing but it should go also it does the computation in a poor way because we should never compute the time of an event from the beginning of the simulation just the interval between events because simulations can be very long — too long to store in a double accurately | 1 |
272,782 | 29,795,087,573 | IssuesEvent | 2023-06-16 01:09:58 | billmcchesney1/flowgate | https://api.github.com/repos/billmcchesney1/flowgate | closed | CVE-2022-0691 (Critical) detected in url-parse-1.4.3.tgz - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-0691 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.3.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.3.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.3.tgz</a></p>
<p>Path to dependency file: /ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.9.
<p>Publish Date: 2022-02-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0691>CVE-2022-0691</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0691">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0691</a></p>
<p>Release Date: 2022-02-21</p>
<p>Fix Resolution: 1.5.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2022-0691 (Critical) detected in url-parse-1.4.3.tgz - autoclosed - ## CVE-2022-0691 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.3.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.3.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.3.tgz</a></p>
<p>Path to dependency file: /ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.9.
<p>Publish Date: 2022-02-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0691>CVE-2022-0691</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0691">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0691</a></p>
<p>Release Date: 2022-02-21</p>
<p>Fix Resolution: 1.5.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_test | cve critical detected in url parse tgz autoclosed cve critical severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file ui package json path to vulnerable library ui node modules url parse package json dependency hierarchy x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue | 0 |
201,809 | 15,225,662,538 | IssuesEvent | 2021-02-18 07:42:48 | WeiXian042901/fyp_repository | https://api.github.com/repos/WeiXian042901/fyp_repository | opened | FU_029_Quiz Play Page(No Answer Selected) | Acceptance Test Quiz User | **Test Scenario**
- User does not select any of the answer options before the timer ended
**Test Case**
- Check that the the correct answer and wrong answers for that question is revealed after the timer ends
**Pre-Conditions**
- User has successfully entered the application
- User clicked on the “Quizzes” Option
- User selected the “Testing title(With Description)” quiz option
- User clicked on the “Start Quiz” button.
**Test-Steps**
**Test Data**
**Expected Results**
- The user should be informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red
**Actual Results**
- The user is informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red
**Pass/Fail**
- Pass
**Date Tested**
- 10th February 2021
**Tested By**
- Zachary Tan | 1.0 | FU_029_Quiz Play Page(No Answer Selected) - **Test Scenario**
- User does not select any of the answer options before the timer ended
**Test Case**
- Check that the the correct answer and wrong answers for that question is revealed after the timer ends
**Pre-Conditions**
- User has successfully entered the application
- User clicked on the “Quizzes” Option
- User selected the “Testing title(With Description)” quiz option
- User clicked on the “Start Quiz” button.
**Test-Steps**
**Test Data**
**Expected Results**
- The user should be informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red
**Actual Results**
- The user is informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red
**Pass/Fail**
- Pass
**Date Tested**
- 10th February 2021
**Tested By**
- Zachary Tan | test | fu quiz play page no answer selected test scenario user does not select any of the answer options before the timer ended test case check that the the correct answer and wrong answers for that question is revealed after the timer ends pre conditions user has successfully entered the application user clicked on the “quizzes” option user selected the “testing title with description ” quiz option user clicked on the “start quiz” button test steps test data expected results the user should be informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red actual results the user is informed of the correct answer with it’s placeholder highlighted green and the wrong answers with all it’s placeholders highlighted red pass fail pass date tested february tested by zachary tan | 1 |
321,067 | 27,503,802,831 | IssuesEvent | 2023-03-06 00:05:04 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix math.test_tensorflow_ceil | TensorFlow Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4327872357/jobs/7556991307" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4327872357/jobs/7556992184" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4011265503/jobs/6888662116" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>Not found</summary>
Not found
</details>
| 1.0 | Fix math.test_tensorflow_ceil - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4327872357/jobs/7556991307" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4327872357/jobs/7556992184" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4011265503/jobs/6888662116" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>Not found</summary>
Not found
</details>
| test | fix math test tensorflow ceil tensorflow img src torch img src numpy img src jax img src not found not found | 1 |
66,202 | 6,992,818,101 | IssuesEvent | 2017-12-15 08:51:38 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Agent will stop and panic after start a while | area/agent area/windows kind/bug status/resolved status/to-test team/cn | **Rancher versions:**
rancher/server: v1.6.13-rc2
rancher/agent:
**Infrastructure Stack versions:**
healthcheck:
ipsec:
network-services:
scheduler:
kubernetes (if applicable):
**Docker version: (`docker version`,`docker info` preferred)**
**Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)**
**Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)**
**Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
**Environment Template: (Cattle/Kubernetes/Swarm/Mesos)**
Windows
**Steps to Reproduce:**
1. In Windows environment, Add host.
2. The agent service will start and print log.
3. after a while, agent print panic log as following
```
fatal error: unexpected signal during runtime execution
[signal 0xc0000005 code=0x0 addr=0x1000002b8 pc=0x40db92]
goroutine 29 [running]:
runtime.throw(0xa103fb, 0x2a)
/usr/local/go/src/runtime/panic.go:596 +0x9c fp=0xc0422dbe18 sp=0xc0422dbdf8
runtime.sigpanic()
/usr/local/go/src/runtime/signal_windows.go:155 +0x18b fp=0xc0422dbe48 sp=0xc0422dbe18
runtime.unlock(0xc9a978)
/usr/local/go/src/runtime/lock_sema.go:107 +0x72 fp=0xc0422dbe70 sp=0xc0422dbe48
runtime.chansend(0x8c6c20, 0xc9a920, 0xc0422dbfb0, 0xc0422dbf01, 0x71824a, 0xc0422dbf9c)
/usr/local/go/src/runtime/chan.go:179 +0x8af fp=0xc0422dbf20 sp=0xc0422dbe70
runtime.chansend1(0x8c6c20, 0xc9a920, 0xc0422dbfb0)
/usr/local/go/src/runtime/chan.go:113 +0x4d fp=0xc0422dbf60 sp=0xc0422dbf20
github.com/rancher/agent/vendor/github.com/Microsoft/go-winio.ioCompletionProcessor(0x270)
/go/src/github.com/rancher/agent/vendor/github.com/Microsoft/go-winio/file.go:131 +0xea fp=0xc0422dbfd8 sp=0xc0422dbf60
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:2197 +0x1 fp=0xc0422dbfe0 sp=0xc0422dbfd8
created by github.com/rancher/agent/vendor/github.com/Microsoft/go-winio.initIo
/go/src/github.com/rancher/agent/vendor/github.com/Microsoft/go-winio/file.go:55 +0x87
goroutine 1 [IO wait]:
net.runtime_pollWait(0xe5cee0, 0x72, 0xc98350)
/usr/local/go/src/runtime/netpoll.go:164 +0x60
net.(*pollDesc).wait(0xc0420b9be0, 0x72, 0xc98350, 0x0)
/usr/local/go/src/net/fd_poll_runtime.go:75 +0x3f
net.(*ioSrv).ExecIO(0xc0420044a0, 0xc0420b9aa0, 0x9fd420, 0x7, 0xa18d18, 0x2, 0x0, 0x0)
/usr/local/go/src/net/fd_windows.go:196 +0xfe
net.(*netFD).Read(0xc0420b9a40, 0xc042468000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/fd_windows.go:452 +0x13d
net.(*conn).Read(0xc042005af8, 0xc042468000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:181 +0x77
bufio.(*Reader).Read(0xc0423fcfc0, 0xc04221ade0, 0x2, 0x8, 0xc0424db8b0, 0x40faef, 0x10)
/usr/local/go/src/bufio/bufio.go:213 +0x319
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).readFull(0xc0423dcc00, 0xc04221ade0, 0x2, 0x8, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:617 +0x98
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).advanceFrame(0xc0423dcc00, 0xa, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:641 +0x96
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).NextReader(0xc0423dcc00, 0xc04238ea80, 0x89a1c0, 0x16, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:780 +0x5e
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).ReadMessage(0xc0423dcc00, 0xc0424a41a0, 0xc04245cf90, 0xc04238ea80, 0xc042005dd8, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:852 +0x36
github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events.(*EventRouter).run(0xc04225dea0, 0xc9a4a0, 0xc04245e570, 0x0, 0x0, 0x0, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events/listener.go:139 +0x760
github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events.(*EventRouter).StartWithoutCreate(0xc04225dea0, 0x0, 0x0, 0xc04235eea0)
/go/src/github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events/listener.go:93 +0x84
github.com/rancher/agent/events.Listen(0xc04235eea0, 0x1d, 0xc04235eec0, 0x14, 0xc042325fb0, 0x28, 0xfa, 0x0, 0x60)
/go/src/github.com/rancher/agent/events/listener.go:53 +0x3fc
main.main()
/go/src/github.com/rancher/agent/main.go:59 +0x421
...
```
| 1.0 | Agent will stop and panic after start a while - **Rancher versions:**
rancher/server: v1.6.13-rc2
rancher/agent:
**Infrastructure Stack versions:**
healthcheck:
ipsec:
network-services:
scheduler:
kubernetes (if applicable):
**Docker version: (`docker version`,`docker info` preferred)**
**Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)**
**Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)**
**Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
**Environment Template: (Cattle/Kubernetes/Swarm/Mesos)**
Windows
**Steps to Reproduce:**
1. In Windows environment, Add host.
2. The agent service will start and print log.
3. after a while, agent print panic log as following
```
fatal error: unexpected signal during runtime execution
[signal 0xc0000005 code=0x0 addr=0x1000002b8 pc=0x40db92]
goroutine 29 [running]:
runtime.throw(0xa103fb, 0x2a)
/usr/local/go/src/runtime/panic.go:596 +0x9c fp=0xc0422dbe18 sp=0xc0422dbdf8
runtime.sigpanic()
/usr/local/go/src/runtime/signal_windows.go:155 +0x18b fp=0xc0422dbe48 sp=0xc0422dbe18
runtime.unlock(0xc9a978)
/usr/local/go/src/runtime/lock_sema.go:107 +0x72 fp=0xc0422dbe70 sp=0xc0422dbe48
runtime.chansend(0x8c6c20, 0xc9a920, 0xc0422dbfb0, 0xc0422dbf01, 0x71824a, 0xc0422dbf9c)
/usr/local/go/src/runtime/chan.go:179 +0x8af fp=0xc0422dbf20 sp=0xc0422dbe70
runtime.chansend1(0x8c6c20, 0xc9a920, 0xc0422dbfb0)
/usr/local/go/src/runtime/chan.go:113 +0x4d fp=0xc0422dbf60 sp=0xc0422dbf20
github.com/rancher/agent/vendor/github.com/Microsoft/go-winio.ioCompletionProcessor(0x270)
/go/src/github.com/rancher/agent/vendor/github.com/Microsoft/go-winio/file.go:131 +0xea fp=0xc0422dbfd8 sp=0xc0422dbf60
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:2197 +0x1 fp=0xc0422dbfe0 sp=0xc0422dbfd8
created by github.com/rancher/agent/vendor/github.com/Microsoft/go-winio.initIo
/go/src/github.com/rancher/agent/vendor/github.com/Microsoft/go-winio/file.go:55 +0x87
goroutine 1 [IO wait]:
net.runtime_pollWait(0xe5cee0, 0x72, 0xc98350)
/usr/local/go/src/runtime/netpoll.go:164 +0x60
net.(*pollDesc).wait(0xc0420b9be0, 0x72, 0xc98350, 0x0)
/usr/local/go/src/net/fd_poll_runtime.go:75 +0x3f
net.(*ioSrv).ExecIO(0xc0420044a0, 0xc0420b9aa0, 0x9fd420, 0x7, 0xa18d18, 0x2, 0x0, 0x0)
/usr/local/go/src/net/fd_windows.go:196 +0xfe
net.(*netFD).Read(0xc0420b9a40, 0xc042468000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/fd_windows.go:452 +0x13d
net.(*conn).Read(0xc042005af8, 0xc042468000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/usr/local/go/src/net/net.go:181 +0x77
bufio.(*Reader).Read(0xc0423fcfc0, 0xc04221ade0, 0x2, 0x8, 0xc0424db8b0, 0x40faef, 0x10)
/usr/local/go/src/bufio/bufio.go:213 +0x319
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).readFull(0xc0423dcc00, 0xc04221ade0, 0x2, 0x8, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:617 +0x98
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).advanceFrame(0xc0423dcc00, 0xa, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:641 +0x96
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).NextReader(0xc0423dcc00, 0xc04238ea80, 0x89a1c0, 0x16, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:780 +0x5e
github.com/rancher/agent/vendor/github.com/gorilla/websocket.(*Conn).ReadMessage(0xc0423dcc00, 0xc0424a41a0, 0xc04245cf90, 0xc04238ea80, 0xc042005dd8, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/gorilla/websocket/conn.go:852 +0x36
github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events.(*EventRouter).run(0xc04225dea0, 0xc9a4a0, 0xc04245e570, 0x0, 0x0, 0x0, 0x0, 0x0)
/go/src/github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events/listener.go:139 +0x760
github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events.(*EventRouter).StartWithoutCreate(0xc04225dea0, 0x0, 0x0, 0xc04235eea0)
/go/src/github.com/rancher/agent/vendor/github.com/rancher/event-subscriber/events/listener.go:93 +0x84
github.com/rancher/agent/events.Listen(0xc04235eea0, 0x1d, 0xc04235eec0, 0x14, 0xc042325fb0, 0x28, 0xfa, 0x0, 0x60)
/go/src/github.com/rancher/agent/events/listener.go:53 +0x3fc
main.main()
/go/src/github.com/rancher/agent/main.go:59 +0x421
...
```
| test | agent will stop and panic after start a while rancher versions rancher server rancher agent infrastructure stack versions healthcheck ipsec network services scheduler kubernetes if applicable docker version docker version docker info preferred operating system and kernel cat etc os release uname r preferred type provider of hosts virtualbox bare metal aws gce do setup details single node rancher vs ha rancher internal db vs external db environment template cattle kubernetes swarm mesos windows steps to reproduce in windows environment add host the agent service will start and print log after a while agent print panic log as following fatal error unexpected signal during runtime execution goroutine runtime throw usr local go src runtime panic go fp sp runtime sigpanic usr local go src runtime signal windows go fp sp runtime unlock usr local go src runtime lock sema go fp sp runtime chansend usr local go src runtime chan go fp sp runtime usr local go src runtime chan go fp sp github com rancher agent vendor github com microsoft go winio iocompletionprocessor go src github com rancher agent vendor github com microsoft go winio file go fp sp runtime goexit usr local go src runtime asm s fp sp created by github com rancher agent vendor github com microsoft go winio initio go src github com rancher agent vendor github com microsoft go winio file go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net iosrv execio usr local go src net fd windows go net netfd read usr local go src net fd windows go net conn read usr local go src net net go bufio reader read usr local go src bufio bufio go github com rancher agent vendor github com gorilla websocket conn readfull go src github com rancher agent vendor github com gorilla websocket conn go github com rancher agent vendor github com gorilla websocket conn advanceframe go src github com rancher agent vendor github com gorilla websocket conn go github com rancher agent vendor github com gorilla websocket conn nextreader go src github com rancher agent vendor github com gorilla websocket conn go github com rancher agent vendor github com gorilla websocket conn readmessage go src github com rancher agent vendor github com gorilla websocket conn go github com rancher agent vendor github com rancher event subscriber events eventrouter run go src github com rancher agent vendor github com rancher event subscriber events listener go github com rancher agent vendor github com rancher event subscriber events eventrouter startwithoutcreate go src github com rancher agent vendor github com rancher event subscriber events listener go github com rancher agent events listen go src github com rancher agent events listener go main main go src github com rancher agent main go | 1 |
174,440 | 13,489,302,021 | IssuesEvent | 2020-09-11 13:42:11 | anamnavi/PowerShellGet | https://api.github.com/repos/anamnavi/PowerShellGet | opened | Role Capability resource - test Prerelease parameter | ask testing | need to create Role Capability resource that has preview version and add test
for that param in the pester test file for it.
Todo: ask Amber how to publish pkg to PoshTestGallery | 1.0 | Role Capability resource - test Prerelease parameter - need to create Role Capability resource that has preview version and add test
for that param in the pester test file for it.
Todo: ask Amber how to publish pkg to PoshTestGallery | test | role capability resource test prerelease parameter need to create role capability resource that has preview version and add test for that param in the pester test file for it todo ask amber how to publish pkg to poshtestgallery | 1 |
207,051 | 15,789,466,667 | IssuesEvent | 2021-04-01 22:50:26 | joncampbell123/dosbox-x | https://api.github.com/repos/joncampbell123/dosbox-x | closed | Test program list/requests (2018 December to 2019 Febrary) | Automated testing help wanted | **Is your feature request related to a problem? Please describe.**
I'm going to slow development of DOSBox-X for a bit (leave it to bug fixes) and focus on developing DOS programs to run both within DOSBox-X and on real hardware as a way to validate that emulation is as accurate as possible.
You may remember in July 2018 when I wrote a program to run through the various INT 10h video modes to test that the modes acted as expected and that various basic tweaks to the VGA hardware acted as expected.
The idea is to do the same with other parts of the hardware, though not yet as comprehensive.
These test programs will exist for both IBM PC and NEC PC-98 platforms.
I have my own list to start with, anyone else here is welcome to suggest additional tests.
Some tests will be written to talk to hardware in the same manner as some games or programs are observed to do. | 1.0 | Test program list/requests (2018 December to 2019 Febrary) - **Is your feature request related to a problem? Please describe.**
I'm going to slow development of DOSBox-X for a bit (leave it to bug fixes) and focus on developing DOS programs to run both within DOSBox-X and on real hardware as a way to validate that emulation is as accurate as possible.
You may remember in July 2018 when I wrote a program to run through the various INT 10h video modes to test that the modes acted as expected and that various basic tweaks to the VGA hardware acted as expected.
The idea is to do the same with other parts of the hardware, though not yet as comprehensive.
These test programs will exist for both IBM PC and NEC PC-98 platforms.
I have my own list to start with, anyone else here is welcome to suggest additional tests.
Some tests will be written to talk to hardware in the same manner as some games or programs are observed to do. | test | test program list requests december to febrary is your feature request related to a problem please describe i m going to slow development of dosbox x for a bit leave it to bug fixes and focus on developing dos programs to run both within dosbox x and on real hardware as a way to validate that emulation is as accurate as possible you may remember in july when i wrote a program to run through the various int video modes to test that the modes acted as expected and that various basic tweaks to the vga hardware acted as expected the idea is to do the same with other parts of the hardware though not yet as comprehensive these test programs will exist for both ibm pc and nec pc platforms i have my own list to start with anyone else here is welcome to suggest additional tests some tests will be written to talk to hardware in the same manner as some games or programs are observed to do | 1 |
70,805 | 7,201,197,447 | IssuesEvent | 2018-02-05 21:42:14 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | core-agent-linux v4.0.16 (r4.0) | r4.0-jessie-cur-test r4.0-stretch-cur-test | Update of core-agent-linux to v4.0.16 for Qubes r4.0, see comments below for details.
Built from: https://github.com/QubesOS/qubes-core-agent-linux/commit/1b774f9a873264b1199cdc400d0241c33e36375a
[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.0.15...v4.0.16):
QubesOS/qubes-core-agent-linux@1b774f9 version 4.0.16
QubesOS/qubes-core-agent-linux@7ecb74a Disable automatic scaling in GNOME/GTK applications
QubesOS/qubes-core-agent-linux@4cd16a2 Enable gnome settings daemon xsettings plugin
QubesOS/qubes-core-agent-linux@d4f6eb1 Install KDE actions for KDE5
QubesOS/qubes-core-agent-linux@7d82029 Fix kdialog --progressbar usage
QubesOS/qubes-core-agent-linux@a8aa41e Merge remote-tracking branch 'qubesos/pr/83'
QubesOS/qubes-core-agent-linux@4c47ce1 qrexec: fix infinite loop when multiple services are waiting for GUI
QubesOS/qubes-core-agent-linux@c324b16 firewall: allow also related traffic
QubesOS/qubes-core-agent-linux@3a83623 firewall: don't crash the whole qubes-firewall service on DNS fail
Referenced issues:
QubesOS/qubes-issues#3449
QubesOS/qubes-issues#3108
QubesOS/qubes-issues#3277
QubesOS/qubes-issues#3406
QubesOS/qubes-issues#3433
If you're release manager, you can issue GPG-inline signed command:
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 current repo` (available 7 days from now)
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| 2.0 | core-agent-linux v4.0.16 (r4.0) - Update of core-agent-linux to v4.0.16 for Qubes r4.0, see comments below for details.
Built from: https://github.com/QubesOS/qubes-core-agent-linux/commit/1b774f9a873264b1199cdc400d0241c33e36375a
[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.0.15...v4.0.16):
QubesOS/qubes-core-agent-linux@1b774f9 version 4.0.16
QubesOS/qubes-core-agent-linux@7ecb74a Disable automatic scaling in GNOME/GTK applications
QubesOS/qubes-core-agent-linux@4cd16a2 Enable gnome settings daemon xsettings plugin
QubesOS/qubes-core-agent-linux@d4f6eb1 Install KDE actions for KDE5
QubesOS/qubes-core-agent-linux@7d82029 Fix kdialog --progressbar usage
QubesOS/qubes-core-agent-linux@a8aa41e Merge remote-tracking branch 'qubesos/pr/83'
QubesOS/qubes-core-agent-linux@4c47ce1 qrexec: fix infinite loop when multiple services are waiting for GUI
QubesOS/qubes-core-agent-linux@c324b16 firewall: allow also related traffic
QubesOS/qubes-core-agent-linux@3a83623 firewall: don't crash the whole qubes-firewall service on DNS fail
Referenced issues:
QubesOS/qubes-issues#3449
QubesOS/qubes-issues#3108
QubesOS/qubes-issues#3277
QubesOS/qubes-issues#3406
QubesOS/qubes-issues#3433
If you're release manager, you can issue GPG-inline signed command:
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 current repo` (available 7 days from now)
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload core-agent-linux 1b774f9a873264b1199cdc400d0241c33e36375a r4.0 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| test | core agent linux update of core agent linux to for qubes see comments below for details built from qubesos qubes core agent linux version qubesos qubes core agent linux disable automatic scaling in gnome gtk applications qubesos qubes core agent linux enable gnome settings daemon xsettings plugin qubesos qubes core agent linux install kde actions for qubesos qubes core agent linux fix kdialog progressbar usage qubesos qubes core agent linux merge remote tracking branch qubesos pr qubesos qubes core agent linux qrexec fix infinite loop when multiple services are waiting for gui qubesos qubes core agent linux firewall allow also related traffic qubesos qubes core agent linux firewall don t crash the whole qubes firewall service on dns fail referenced issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload core agent linux current repo available days from now upload core agent linux current dists repo you can choose subset of distributions like vm vm available days from now upload core agent linux security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it | 1 |
39,856 | 5,252,864,773 | IssuesEvent | 2017-02-02 07:08:53 | dccarter/sparc | https://api.github.com/repos/dccarter/sparc | opened | System testing framework | testing | Sparc needs to be system tested in order to catch some bugs in early development. We need to develop (or adopt) a system test framework that will be used to test this web framework and its features. | 1.0 | System testing framework - Sparc needs to be system tested in order to catch some bugs in early development. We need to develop (or adopt) a system test framework that will be used to test this web framework and its features. | test | system testing framework sparc needs to be system tested in order to catch some bugs in early development we need to develop or adopt a system test framework that will be used to test this web framework and its features | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.