Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
7,593
10,704,514,124
IssuesEvent
2019-10-24 11:55:07
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
obsoletion GO:0075171 and similar in response to host branch
multi-species process
regulation of symbiont pathways in response to host should probably be obsoleted if not used. for example GO:0075171 regulation of MAP kinase-mediated signal transduction in response to host No doubt the MAP kinase stress pathways are activated in response to host defenses but I don't think we have any examples yet (and we could use the regulation of MAP kinase-mediated signal transduction and precompose the "response to host" part if we did)
1.0
obsoletion GO:0075171 and similar in response to host branch - regulation of symbiont pathways in response to host should probably be obsoleted if not used. for example GO:0075171 regulation of MAP kinase-mediated signal transduction in response to host No doubt the MAP kinase stress pathways are activated in response to host defenses but I don't think we have any examples yet (and we could use the regulation of MAP kinase-mediated signal transduction and precompose the "response to host" part if we did)
process
obsoletion go and similar in response to host branch regulation of symbiont pathways in response to host should probably be obsoleted if not used for example go regulation of map kinase mediated signal transduction in response to host no doubt the map kinase stress pathways are activated in response to host defenses but i don t think we have any examples yet and we could use the regulation of map kinase mediated signal transduction and precompose the response to host part if we did
1
11,633
14,492,063,132
IssuesEvent
2020-12-11 06:10:51
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Kubeflow v1.2 release
area/1.2.0 area/docs community/maintenance effort/2-weeks+ kind/feature kind/process platform/aws priority/p1
/kind process Time line: - RC Release: Nov 7 - Official Release: Nov 16 WGs and projects Area | release czar | Tracking Issue --- | --- | --- centraldashboard | | https://github.com/kubeflow/kubeflow/pull/5412 docs |@RFMVasconcelos | kubeflow/website#2322 fairing | @jinchihe | | feast | @woop | | katib | @andreyvelich | https://github.com/kubeflow/manifests/pull/1593 | kfctl | @Jeffwan | https://github.com/kubeflow/kfctl/issues/421 | kfserving | @yuzisun @animeshsingh @cliveseldon | https://github.com/kubeflow/manifests/pull/1600 https://github.com/kubeflow/manifests/pull/1575 | manifests | @Jeffwan | kubeflow/manifests#1597 metadata | | notebooks | @kimwnasptd , @jtfogarty | | pipelines | @Bobgy | https://github.com/kubeflow/manifests/pull/1605 | training | @johnugeorge @gaocegege @terrytangyuan | | Platforms Platforms | release czar | Tracking Issue --- | --- | --- aws | @PatrickXYS | | azure | @aronchick | | ibm | @animeshsingh @adrian555 @shawnzhu | | gcp | @jlewi @Bobgy | | minikf | @vkoukis | |
1.0
Kubeflow v1.2 release - /kind process Time line: - RC Release: Nov 7 - Official Release: Nov 16 WGs and projects Area | release czar | Tracking Issue --- | --- | --- centraldashboard | | https://github.com/kubeflow/kubeflow/pull/5412 docs |@RFMVasconcelos | kubeflow/website#2322 fairing | @jinchihe | | feast | @woop | | katib | @andreyvelich | https://github.com/kubeflow/manifests/pull/1593 | kfctl | @Jeffwan | https://github.com/kubeflow/kfctl/issues/421 | kfserving | @yuzisun @animeshsingh @cliveseldon | https://github.com/kubeflow/manifests/pull/1600 https://github.com/kubeflow/manifests/pull/1575 | manifests | @Jeffwan | kubeflow/manifests#1597 metadata | | notebooks | @kimwnasptd , @jtfogarty | | pipelines | @Bobgy | https://github.com/kubeflow/manifests/pull/1605 | training | @johnugeorge @gaocegege @terrytangyuan | | Platforms Platforms | release czar | Tracking Issue --- | --- | --- aws | @PatrickXYS | | azure | @aronchick | | ibm | @animeshsingh @adrian555 @shawnzhu | | gcp | @jlewi @Bobgy | | minikf | @vkoukis | |
process
kubeflow release kind process time line rc release nov official release nov wgs and projects area release czar tracking issue centraldashboard docs rfmvasconcelos kubeflow website fairing jinchihe feast woop katib andreyvelich kfctl jeffwan kfserving yuzisun animeshsingh cliveseldon manifests jeffwan kubeflow manifests metadata notebooks kimwnasptd jtfogarty pipelines bobgy training johnugeorge gaocegege terrytangyuan platforms platforms release czar tracking issue aws patrickxys azure aronchick ibm animeshsingh shawnzhu gcp jlewi bobgy minikf vkoukis
1
123,309
10,263,664,907
IssuesEvent
2019-08-22 14:47:33
coala/coala-bears
https://api.github.com/repos/coala/coala-bears
opened
HgCommitBearTest: BitBucket hosting of hg repos ends June 1 2020
area/bears area/genericbears area/tests
ERROR: type should be string, got "https://bitbucket.org/blog/sunsetting-mercurial-support-in-bitbucket\r\n\r\nIf another mercurial host can be used for tests, that is an option, otherwise we can disable the tests and lower the required coverage to 0% (`.ci/generate_coverage_thresholds.py`), or....\r\n\r\nKill the bear!\r\n\r\nlabel: un-bear-proposal"
1.0
HgCommitBearTest: BitBucket hosting of hg repos ends June 1 2020 - https://bitbucket.org/blog/sunsetting-mercurial-support-in-bitbucket If another mercurial host can be used for tests, that is an option, otherwise we can disable the tests and lower the required coverage to 0% (`.ci/generate_coverage_thresholds.py`), or.... Kill the bear! label: un-bear-proposal
non_process
hgcommitbeartest bitbucket hosting of hg repos ends june if another mercurial host can be used for tests that is an option otherwise we can disable the tests and lower the required coverage to ci generate coverage thresholds py or kill the bear label un bear proposal
0
11,396
14,232,928,475
IssuesEvent
2020-11-18 11:23:13
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
prisma init: add .env to root instead of ./prisma
kind/improvement process/candidate team/client tech/typescript
The `prisma init` command should add the `.env` file to the root of the project instead of the `prisma` dir. This needs conflict resolution logic in case there already is a `.env` file: - If there is no conflict with the existing env vars, append - If the env var about to be added already is defined, throw an error
1.0
prisma init: add .env to root instead of ./prisma - The `prisma init` command should add the `.env` file to the root of the project instead of the `prisma` dir. This needs conflict resolution logic in case there already is a `.env` file: - If there is no conflict with the existing env vars, append - If the env var about to be added already is defined, throw an error
process
prisma init add env to root instead of prisma the prisma init command should add the env file to the root of the project instead of the prisma dir this needs conflict resolution logic in case there already is a env file if there is no conflict with the existing env vars append if the env var about to be added already is defined throw an error
1
12,998
15,359,722,888
IssuesEvent
2021-03-01 16:09:19
Psychoanalytic-Electronic-Publishing/OpenPubArchive-Content-Server
https://api.github.com/repos/Psychoanalytic-Electronic-Publishing/OpenPubArchive-Content-Server
closed
Solrpy library still in use
feature in process
Most Solr functions have been transitioned to Pysolr...but a few remain using Solrpy. Need to transition those and eliminate Solrpy, just for code minimization and cleaner design.
1.0
Solrpy library still in use - Most Solr functions have been transitioned to Pysolr...but a few remain using Solrpy. Need to transition those and eliminate Solrpy, just for code minimization and cleaner design.
process
solrpy library still in use most solr functions have been transitioned to pysolr but a few remain using solrpy need to transition those and eliminate solrpy just for code minimization and cleaner design
1
11,362
14,175,715,420
IssuesEvent
2020-11-12 22:05:22
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Select pipeline resource version
Pri1 devops-cicd-process/tech devops/prod doc-bug
Is there a way to select the pipeline resource version when manually running a pipeline through the UI? We have a large app with a long enough build time to necessitate breaking out the build and release into two separate pipelines. Without a way to specify / verify the released version, it is causing us to go back to the classic releases. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=example) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Select pipeline resource version - Is there a way to select the pipeline resource version when manually running a pipeline through the UI? We have a large app with a long enough build time to necessitate breaking out the build and release into two separate pipelines. Without a way to specify / verify the released version, it is causing us to go back to the classic releases. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=example) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
select pipeline resource version is there a way to select the pipeline resource version when manually running a pipeline through the ui we have a large app with a long enough build time to necessitate breaking out the build and release into two separate pipelines without a way to specify verify the released version it is causing us to go back to the classic releases document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
7,726
10,833,433,186
IssuesEvent
2019-11-11 12:55:25
codeuniversity/smag-mvp
https://api.github.com/repos/codeuniversity/smag-mvp
opened
Index face encoding changes into elastic search
Image Processing
- read changes from postgres - index into elastic search (needs addon, linked in epic)
1.0
Index face encoding changes into elastic search - - read changes from postgres - index into elastic search (needs addon, linked in epic)
process
index face encoding changes into elastic search read changes from postgres index into elastic search needs addon linked in epic
1
693
3,184,428,435
IssuesEvent
2015-09-27 10:55:42
bskinn/opan
https://api.github.com/repos/bskinn/opan
opened
Roll out metaclassed iteration on OPANError and Enum classes
enhancement in process
class __metaclass__(type): # Enable iteration over variables containing their own name def __iter__(self): for item in self.__dict__: if item == self.__dict__[item]: yield item Developing in branch `iterable-class`. Implement `OPANError` changes where relevant (cull `typecodes` objects if feasible). Create `OPANEnum` superclass in `const.py`, pull all `E_`, etc. classes under it, and roll out same enhancement for enums, possibly eliminating the `E` member.
1.0
Roll out metaclassed iteration on OPANError and Enum classes - class __metaclass__(type): # Enable iteration over variables containing their own name def __iter__(self): for item in self.__dict__: if item == self.__dict__[item]: yield item Developing in branch `iterable-class`. Implement `OPANError` changes where relevant (cull `typecodes` objects if feasible). Create `OPANEnum` superclass in `const.py`, pull all `E_`, etc. classes under it, and roll out same enhancement for enums, possibly eliminating the `E` member.
process
roll out metaclassed iteration on opanerror and enum classes class metaclass type enable iteration over variables containing their own name def iter self for item in self dict if item self dict yield item developing in branch iterable class implement opanerror changes where relevant cull typecodes objects if feasible create opanenum superclass in const py pull all e etc classes under it and roll out same enhancement for enums possibly eliminating the e member
1
280,692
8,685,268,319
IssuesEvent
2018-12-03 07:01:25
pulumi/docs
https://api.github.com/repos/pulumi/docs
closed
Eliminate @cloud from all AWS-specific Tutorials
area/docs priority/P1
Now that we have advanced functionality in the AWS package itself, we should be able to convert all AWS-specific Tutorials so that they just use `@pulumi/aws`. The one I'm unsure about is the containers example, but I would love to see us make an attempt at that also (and perhaps that would inform the ongoing work we have at the AWS abstraction layer).
1.0
Eliminate @cloud from all AWS-specific Tutorials - Now that we have advanced functionality in the AWS package itself, we should be able to convert all AWS-specific Tutorials so that they just use `@pulumi/aws`. The one I'm unsure about is the containers example, but I would love to see us make an attempt at that also (and perhaps that would inform the ongoing work we have at the AWS abstraction layer).
non_process
eliminate cloud from all aws specific tutorials now that we have advanced functionality in the aws package itself we should be able to convert all aws specific tutorials so that they just use pulumi aws the one i m unsure about is the containers example but i would love to see us make an attempt at that also and perhaps that would inform the ongoing work we have at the aws abstraction layer
0
7,605
10,721,495,920
IssuesEvent
2019-10-27 03:10:06
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
Disable esModuleInterop
type: process
Ideally, we should not need `esModuleInterop` configuration for tsconfig.
1.0
Disable esModuleInterop - Ideally, we should not need `esModuleInterop` configuration for tsconfig.
process
disable esmoduleinterop ideally we should not need esmoduleinterop configuration for tsconfig
1
186,689
14,404,317,455
IssuesEvent
2020-12-03 17:07:59
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
JDK11 JIT Server : cmdLineTester_jvmtitests_hcr_OSRG_nongold_2 ** ASSERTION FAILED ** at jswalk.c:538: ((0 ))
comp:jitserver test failure
Failure link ------------ From an internal build `Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/32/` ``` 23:00:01 openjdk version "11.0.7-internal" 2020-04-14 23:00:01 OpenJDK Runtime Environment (build 11.0.7-internal+0-adhoc.jenkins.BuildJDK11x86-64linuxjitNightly) 23:00:01 Eclipse OpenJ9 VM (build ibm_sdk-0852f80a4a, JRE 11 Linux amd64-64-Bit Compressed References 20200328_33 (JIT enabled, AOT enabled) 23:00:01 OpenJ9 - 0852f80a4a 23:00:01 OMR - e9bed7888 23:00:01 JCL - 34c3dd7d55b based on jdk-11.0.7+9) ``` Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` Testing: rc013 Test start time: 2020/03/28 22:43:41 Pacific Standard Time Running command: "/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdkbinary/j2sdk-image/bin/java" -XX:+UseJITServer -Xcompressedrefs -Xcompressedrefs -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation -Xdump -agentlib:jvmtitest=test:rc013 -cp "/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner Time spent starting: 2 milliseconds Time spent executing: 5671 milliseconds Test result: FAILED Output from test: [ERR] JITServer is currently a technology preview. Its use is not yet supported [ERR] JVMCDRT000E Unable to locate JIT stack map - aborting VM [ERR] JVMCDRT001E Method: java/lang/NullPointerException.<init>(Ljava/lang/String;)V (00000000024821F0) [ERR] JVMCDRT002E Failing PC: 00007F3A98A191E3 (offset 0000000000000023), metaData = 00007F3A924C5278 [ERR] 05:43:41.866 0x244aa00j9codertvm(j9ji.110 * ** ASSERTION FAILED ** at jswalk.c:538: ((0 )) [ERR] JVMDUMP039I Processing dump event "traceassert", detail "" at 2020/03/28 22:43:41 - please wait. [ERR] JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/core.20200328.224341.11863.0001.dmp' in response to an event [ERR] JVMPORT030W /proc/sys/kernel/core_pattern setting "|/usr/libexec/abrt-hook-ccpp %s %c %p %u %g %t e" specifies that the core dump is to be piped to an external program. Attempting to rename either core or core.11879. [ERR] [ERR] JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/core.20200328.224341.11863.0001.dmp [ERR] JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/javacore.20200328.224341.11863.0002.txt' in response to an event [ERR] JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/javacore.20200328.224341.11863.0002.txt [ERR] JVMDUMP032I JVM requested Snap dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/Snap.20200328.224341.11863.0003.trc' in response to an event [ERR] JVMDUMP010I Snap dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/Snap.20200328.224341.11863.0003.trc [ERR] JVMDUMP013I Processed dump event "traceassert", detail "". >> Success condition was not found: [Return code: 0] ``` To rebuild the failed tests in <Jenkins URL>=https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder, use the following links: https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=functional&PLATFORM=x86-64_linux&TARGET=cmdLineTester_jvmtitests_hcr_OSRG_nongold_2
1.0
JDK11 JIT Server : cmdLineTester_jvmtitests_hcr_OSRG_nongold_2 ** ASSERTION FAILED ** at jswalk.c:538: ((0 )) - Failure link ------------ From an internal build `Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/32/` ``` 23:00:01 openjdk version "11.0.7-internal" 2020-04-14 23:00:01 OpenJDK Runtime Environment (build 11.0.7-internal+0-adhoc.jenkins.BuildJDK11x86-64linuxjitNightly) 23:00:01 Eclipse OpenJ9 VM (build ibm_sdk-0852f80a4a, JRE 11 Linux amd64-64-Bit Compressed References 20200328_33 (JIT enabled, AOT enabled) 23:00:01 OpenJ9 - 0852f80a4a 23:00:01 OMR - e9bed7888 23:00:01 JCL - 34c3dd7d55b based on jdk-11.0.7+9) ``` Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` Testing: rc013 Test start time: 2020/03/28 22:43:41 Pacific Standard Time Running command: "/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdkbinary/j2sdk-image/bin/java" -XX:+UseJITServer -Xcompressedrefs -Xcompressedrefs -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation -Xdump -agentlib:jvmtitest=test:rc013 -cp "/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner Time spent starting: 2 milliseconds Time spent executing: 5671 milliseconds Test result: FAILED Output from test: [ERR] JITServer is currently a technology preview. Its use is not yet supported [ERR] JVMCDRT000E Unable to locate JIT stack map - aborting VM [ERR] JVMCDRT001E Method: java/lang/NullPointerException.<init>(Ljava/lang/String;)V (00000000024821F0) [ERR] JVMCDRT002E Failing PC: 00007F3A98A191E3 (offset 0000000000000023), metaData = 00007F3A924C5278 [ERR] 05:43:41.866 0x244aa00j9codertvm(j9ji.110 * ** ASSERTION FAILED ** at jswalk.c:538: ((0 )) [ERR] JVMDUMP039I Processing dump event "traceassert", detail "" at 2020/03/28 22:43:41 - please wait. [ERR] JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/core.20200328.224341.11863.0001.dmp' in response to an event [ERR] JVMPORT030W /proc/sys/kernel/core_pattern setting "|/usr/libexec/abrt-hook-ccpp %s %c %p %u %g %t e" specifies that the core dump is to be piped to an external program. Attempting to rename either core or core.11879. [ERR] [ERR] JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/core.20200328.224341.11863.0001.dmp [ERR] JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/javacore.20200328.224341.11863.0002.txt' in response to an event [ERR] JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/javacore.20200328.224341.11863.0002.txt [ERR] JVMDUMP032I JVM requested Snap dump using '/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/Snap.20200328.224341.11863.0003.trc' in response to an event [ERR] JVMDUMP010I Snap dump written to /home/jenkins/workspace/Test_openjdk11_j9_extended.functional_x86-64_linux_jit_Nightly/openjdk-tests/TKG/test_output_15854512416865/cmdLineTester_jvmtitests_hcr_OSRG_nongold_2/Snap.20200328.224341.11863.0003.trc [ERR] JVMDUMP013I Processed dump event "traceassert", detail "". >> Success condition was not found: [Return code: 0] ``` To rebuild the failed tests in <Jenkins URL>=https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder, use the following links: https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=functional&PLATFORM=x86-64_linux&TARGET=cmdLineTester_jvmtitests_hcr_OSRG_nongold_2
non_process
jit server cmdlinetester jvmtitests hcr osrg nongold assertion failed at jswalk c failure link from an internal build test extended functional linux jit nightly openjdk version internal openjdk runtime environment build internal adhoc jenkins eclipse vm build ibm sdk jre linux bit compressed references jit enabled aot enabled omr jcl based on jdk optional info failure output captured from console output testing test start time pacific standard time running command home jenkins workspace test extended functional linux jit nightly openjdkbinary image bin java xx usejitserver xcompressedrefs xcompressedrefs xgcpolicy optthruput xdebug xrunjdwp transport dt socket address server y onthrow no pkg foo launch echo xjit enableosr enableosronguardfailure count disableasynccompilation xdump agentlib jvmtitest test cp home jenkins workspace test extended functional linux jit nightly openjdk tests tkg jvmtest functional cmdlinetests jvmtitests jvmtitest jar home jenkins workspace test extended functional linux jit nightly openjdk tests tkg tkg lib asm all jar com ibm jvmti tests util testrunner time spent starting milliseconds time spent executing milliseconds test result failed output from test jitserver is currently a technology preview its use is not yet supported unable to locate jit stack map aborting vm method java lang nullpointerexception ljava lang string v failing pc offset metadata assertion failed at jswalk c processing dump event traceassert detail at please wait jvm requested system dump using home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold core dmp in response to an event proc sys kernel core pattern setting usr libexec abrt hook ccpp s c p u g t e specifies that the core dump is to be piped to an external program attempting to rename either core or core system dump written to home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold core dmp jvm requested java dump using home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold javacore txt in response to an event java dump written to home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold javacore txt jvm requested snap dump using home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold snap trc in response to an event snap dump written to home jenkins workspace test extended functional linux jit nightly openjdk tests tkg test output cmdlinetester jvmtitests hcr osrg nongold snap trc processed dump event traceassert detail success condition was not found to rebuild the failed tests in use the following links
0
17,090
22,598,142,335
IssuesEvent
2022-06-29 06:31:26
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
opened
Flaky ReplayStateRandomizedPropertyTest:shouldRestoreStateAtEachStepInExecution
kind/flake team/process-automation
**Summary** Failed in stable/8.0 and stable/1.3 - How often does the test fail? 2 - Does it block your work? No - Do we suspect that it is a real failure? **Failures** > Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins <details><summary>Example assertion failure</summary> <pre> Stacktrace org.awaitility.core.ConditionTimeoutException: Condition with alias 'await that the replay state is equal to the processing state' didn't complete within 10 seconds because assertion condition defined as a io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest Multiple Failures (1 failure) -- failure 1 -- [The state column 'PROCESS_SUBSCRIPTION_BY_KEY' has different entries after replay] Expecting map: {"[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"} to contain only: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] map entries not found: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] and map entries not expected: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"] at ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$3(ReplayStateRandomizedPropertyTest.java:148). at org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:165) at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:119) at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:31) at org.awaitility.core.ConditionFactory.until(ConditionFactory.java:895) at org.awaitility.core.ConditionFactory.untilAsserted(ConditionFactory.java:679) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.stopAndRestartEngineAndCompareStates(ReplayStateRandomizedPropertyTest.java:122) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.shouldRestoreStateAtEachStepInExecution(ReplayStateRandomizedPropertyTest.java:82) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at org.junit.runner.JUnitCore.run(JUnitCore.java:115) at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80) at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.lambda$execute$1(JUnitPlatformProvider.java:199) at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:193) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:120) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: org.assertj.core.error.AssertJMultipleFailuresError: Multiple Failures (1 failure) -- failure 1 -- [The state column 'PROCESS_SUBSCRIPTION_BY_KEY' has different entries after replay] Expecting map: {"[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"} to contain only: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] map entries not found: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] and map entries not expected: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"] at ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$3(ReplayStateRandomizedPropertyTest.java:148) at jdk.internal.reflect.GeneratedConstructorAccessor42.newInstance(Unknown Source) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$4(ReplayStateRandomizedPropertyTest.java:152) at org.awaitility.core.AssertionCondition.lambda$new$0(AssertionCondition.java:53) at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:222) at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:209) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) </pre> </details> **Hypotheses** **Logs** <details><summary>Logs</summary> <pre> Standard Output 16:02:34.308 [Broker-0-LogStream-1] DEBUG io.camunda.zeebe.logstreams - Configured log appender back pressure at partition 1 as AppenderVegasCfg{initialLimit=1024, maxConcurrency=32768, alphaLimit=0.7, betaLimit=0.95}. Window limiting is disabled 16:02:34.492 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:34.492 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@29a4be77)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@27634533, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@235cd507, configuration: Configuration(false)] 16:02:34.494 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:34.495 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@665e4f08)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6a0502b2, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@57f2881c, configuration: Configuration(false)] 16:02:34.496 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@12439b08)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@671b047, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@44a332a9, configuration: Configuration(false)] 16:02:34.497 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7d2f049)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@947b060, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@23fca87, configuration: Configuration(false)] 16:02:34.499 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@5fd49613)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:34.499 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@4f147518)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:34.501 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7be4182a)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:34.501 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@7f39e23)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:34.503 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:34.503 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: -1, lastWrittenPosition: -1] 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:34.504 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:34.505 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:34.505 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:34.899 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Paused processing for partition 1 16:02:34.899 [] INFO io.camunda.zeebe.logstreams - Paused processing for stream stream-1 16:02:34.902 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:34.902 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:34.975 [] INFO io.camunda.zeebe.logstreams - Closed stream stream-1 16:02:35.163 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:35.163 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@82b217f)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6ecde0a, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@2d34a0f, configuration: Configuration(false)] 16:02:35.165 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:35.166 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@47d1d431)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@16a0356d, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@765cb53d, configuration: Configuration(false)] 16:02:35.167 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@6ec70acb)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@27fe2b7e, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@306279b1, configuration: Configuration(false)] 16:02:35.168 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@2304c7dd)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@75c30024, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@23b07039, configuration: Configuration(false)] 16:02:35.170 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@bc3a0f)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:35.171 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@5832fea)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:35.173 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@682ccd5e)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:35.174 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@483d49da)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:35.176 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:35.203 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: 19, lastWrittenPosition: 22] 16:02:35.203 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:35.203 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:35.204 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:35.663 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Paused processing for partition 1 16:02:35.663 [] INFO io.camunda.zeebe.logstreams - Paused processing for stream stream-1 16:02:35.666 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:35.667 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:35.711 [] INFO io.camunda.zeebe.logstreams - Closed stream stream-1 16:02:36.065 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:36.065 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7c6ce4c0)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@10bce2e0, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@4b428d95, configuration: Configuration(false)] 16:02:36.067 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:36.068 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@505b1058)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@4b0aa477, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@3c56f414, configuration: Configuration(false)] 16:02:36.069 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@63a7f82e)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6b76edd2, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@75c1c5d7, configuration: Configuration(false)] 16:02:36.070 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7ebc692f)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@2206985e, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@7e42f9b1, configuration: Configuration(false)] 16:02:36.073 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@5d664d3)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:36.073 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@183e8057)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:36.076 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@eec2f66)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:36.076 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@670ad07b)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:36.079 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:36.106 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: 37, lastWrittenPosition: 39] 16:02:36.106 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:36.106 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:36.107 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:46.290 [] INFO io.camunda.zeebe.test - Test failed, following records were exported: 16:02:46.488 [] INFO io.camunda.zeebe.test - Compact log representation: -------- ['C'ommand/'E'event/'R'ejection] [valueType] [intent] - #[position]->#[source record position] P[partitionId]K[key] - [summary of value] P9K999 - key; #999 - record position; "ID" element/process id; @"elementid"/[P9K999] - element with ID and key Keys are decomposed into partition id and per partition key (e.g. 2251799813685253 -> P1K005). If single partition, the partition is omitted. Long IDs are shortened (e.g. 'startEvent_5d56488e-0570-416c-ba2d-36d2a3acea78' -> 'star..acea78' -------- C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C JOB_BATCH ACTIVATE - #23-> -1 -1 - "id_4_variables" max: 10 E JOB_BATCH ACTIVATED - #24->#23 K10 - "id_4_variables" 1/10 K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) C JOB COMPLETE - #25-> -1 K09 - K09 -1 retries, in <process ?[?]> with variables: {correlationKey=default_correlation_key} E JOB COMPLETED - #26->#25 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> with variables: {correlationKey=default_correlation_key} E PROC_EVNT TRIGGERING - #27->#25 K11 - @"id_4_va..es_task"[K08] in <process K01[K04]> with variables: {correlationKey=default_correlation_key} C PROC_INST COMPLETE - #28->#25 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #29->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E VAR CREATED - #30->#28 K12 - correlationKey->"default_correlation_key" in <process [K04]> E PROC_INST COMPLETED - #31->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #32->#28 K13 - SEQUENCE_FLOW "sequenc..ae31a49" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #33->#28 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #34->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_MSG_SUB CREATING - #35->#33 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #36->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> C MSG_SUB CREATE - #37-> -1 -1 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E MSG_SUB CREATED - #38->#37 K16 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) C PROC_MSG_SUB CREATE - #39-> -1 -1 - "message_id_0" (inter.) @[K14] in <process ?[K04]> (no vars) C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C JOB_BATCH ACTIVATE - #23-> -1 -1 - "id_4_variables" max: 10 E JOB_BATCH ACTIVATED - #24->#23 K10 - "id_4_variables" 1/10 K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) C JOB COMPLETE - #25-> -1 K09 - K09 -1 retries, in <process ?[?]> with variables: {correlationKey=default_correlation_key} E JOB COMPLETED - #26->#25 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> with variables: {correlationKey=default_correlation_key} E PROC_EVNT TRIGGERING - #27->#25 K11 - @"id_4_va..es_task"[K08] in <process K01[K04]> with variables: {correlationKey=default_correlation_key} C PROC_INST COMPLETE - #28->#25 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #29->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E VAR CREATED - #30->#28 K12 - correlationKey->"default_correlation_key" in <process [K04]> E PROC_INST COMPLETED - #31->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #32->#28 K13 - SEQUENCE_FLOW "sequenc..ae31a49" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #33->#28 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #34->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_MSG_SUB CREATING - #35->#33 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #36->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> C MSG_SUB CREATE - #37-> -1 -1 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E MSG_SUB CREATED - #38->#37 K16 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) C PROC_MSG_SUB CREATE - #39-> -1 -1 - "message_id_0" (inter.) @[K14] in <process ?[K04]> (no vars) E PROC_MSG_SUB CREATED - #40->#39 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) -------------- Deployed Processes ---------------------- process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> --------------- Decomposed keys (for debugging) ----------------- -1 <-> -1 K01 <-> 2251799813685249 K02 <-> 2251799813685250 K03 <-> 2251799813685251 K04 <-> 2251799813685252 K05 <-> 2251799813685253 K06 <-> 2251799813685254 K07 <-> 2251799813685255 K08 <-> 2251799813685256 K09 <-> 2251799813685257 K10 <-> 2251799813685258 K11 <-> 2251799813685259 K12 <-> 2251799813685260 K13 <-> 2251799813685261 K14 <-> 2251799813685262 K15 <-> 2251799813685263 K16 <-> 2251799813685264 16:02:46.495 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:46.496 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:46.590 [Broker-0-LogStream-1] INFO io.camunda.zeebe.logstreams - Close appender for log stream stream-1 16:02:46.591 [stream-1-write-buffer] DEBUG io.camunda.zeebe.dispatcher - Dispatcher closed 16:02:46.591 [Broker-0-LogStream-1] INFO io.camunda.zeebe.logstreams - On closing logstream stream-1 close 6 readers 16:02:46.602 [] DEBUG io.camunda.zeebe.broker.test - Clean up test files on path /tmp/junit12212456625141852810 16:02:46.603 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers' 16:02:46.603 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-actors' 16:02:46.606 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers': closed successfully 16:02:46.606 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-actors': closed successfully 16:02:46.612 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Data of failed test case: TestDataRecord{processSeed=884380440262060521, executionPathSeed=2777154302604920046} 16:02:46.688 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Test case failed at: ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_4_variables_task,jobType=id_4_variables,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} 16:02:46.691 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Execution path of failed test case: ScheduledExecutionStep{step=StepTriggerTimerStartEvent[timeToAdd=PT26M,variables={}], activationTime=PT0S, startTime=PT0S, endTime=PT26M, activationDuration=PT26M} --failed here--> ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_4_variables_task,jobType=id_4_variables,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepPublishMessage[correlationKeyValue=default_correlation_key,messageName=message_id_0,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepPublishMessage[correlationKeyValue=default_correlation_key,messageName=message_id_1,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepActivateBPMNElement[elementId=id_2,variables={}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepActivateAndTimeoutJob[elementId=id_2,jobType=job_id_2,variables={}], activationTime=PT26M, startTime=PT26M, endTime=PT1H26M, activationDuration=PT1H} ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_2,jobType=job_id_2,variables={}], activationTime=PT26M, startTime=PT1H26M, endTime=PT1H26M, activationDuration=PT1H} </pre> </details>
1.0
Flaky ReplayStateRandomizedPropertyTest:shouldRestoreStateAtEachStepInExecution - **Summary** Failed in stable/8.0 and stable/1.3 - How often does the test fail? 2 - Does it block your work? No - Do we suspect that it is a real failure? **Failures** > Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins <details><summary>Example assertion failure</summary> <pre> Stacktrace org.awaitility.core.ConditionTimeoutException: Condition with alias 'await that the replay state is equal to the processing state' didn't complete within 10 seconds because assertion condition defined as a io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest Multiple Failures (1 failure) -- failure 1 -- [The state column 'PROCESS_SUBSCRIPTION_BY_KEY' has different entries after replay] Expecting map: {"[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"} to contain only: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] map entries not found: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] and map entries not expected: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"] at ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$3(ReplayStateRandomizedPropertyTest.java:148). at org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:165) at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:119) at org.awaitility.core.AssertionCondition.await(AssertionCondition.java:31) at org.awaitility.core.ConditionFactory.until(ConditionFactory.java:895) at org.awaitility.core.ConditionFactory.untilAsserted(ConditionFactory.java:679) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.stopAndRestartEngineAndCompareStates(ReplayStateRandomizedPropertyTest.java:122) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.shouldRestoreStateAtEachStepInExecution(ReplayStateRandomizedPropertyTest.java:82) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at org.junit.runner.JUnitCore.run(JUnitCore.java:115) at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80) at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.lambda$execute$1(JUnitPlatformProvider.java:199) at java.base/java.util.Iterator.forEachRemaining(Iterator.java:133) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:193) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:120) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: org.assertj.core.error.AssertJMultipleFailuresError: Multiple Failures (1 failure) -- failure 1 -- [The state column 'PROCESS_SUBSCRIPTION_BY_KEY' has different entries after replay] Expecting map: {"[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"} to contain only: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] map entries not found: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENING","key":2251799813685263}"] and map entries not expected: ["[0, 8, 0, 0, 0, 0, 0, 14, 0, 0, 0, 12, 109, 101, 115, 115, 97, 103, 101, 95, 105, 100, 95, 48]"="{"record":{"subscriptionPartitionId":1,"processInstanceKey":2251799813685252,"elementInstanceKey":2251799813685262,"messageKey":-1,"messageName":"message_id_0","variables":"gA==","interrupting":true,"bpmnProcessId":"process_id_3","correlationKey":"default_correlation_key","elementId":"id_0"},"state":"STATE_OPENED","key":2251799813685263}"] at ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$3(ReplayStateRandomizedPropertyTest.java:148) at jdk.internal.reflect.GeneratedConstructorAccessor42.newInstance(Unknown Source) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) at io.camunda.zeebe.engine.processing.streamprocessor.ReplayStateRandomizedPropertyTest.lambda$stopAndRestartEngineAndCompareStates$4(ReplayStateRandomizedPropertyTest.java:152) at org.awaitility.core.AssertionCondition.lambda$new$0(AssertionCondition.java:53) at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:222) at org.awaitility.core.ConditionAwaiter$ConditionPoller.call(ConditionAwaiter.java:209) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) </pre> </details> **Hypotheses** **Logs** <details><summary>Logs</summary> <pre> Standard Output 16:02:34.308 [Broker-0-LogStream-1] DEBUG io.camunda.zeebe.logstreams - Configured log appender back pressure at partition 1 as AppenderVegasCfg{initialLimit=1024, maxConcurrency=32768, alphaLimit=0.7, betaLimit=0.95}. Window limiting is disabled 16:02:34.492 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:34.492 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@29a4be77)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@27634533, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@235cd507, configuration: Configuration(false)] 16:02:34.494 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:34.495 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@665e4f08)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6a0502b2, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@57f2881c, configuration: Configuration(false)] 16:02:34.496 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@12439b08)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@671b047, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@44a332a9, configuration: Configuration(false)] 16:02:34.497 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7d2f049)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@947b060, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@23fca87, configuration: Configuration(false)] 16:02:34.499 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@5fd49613)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:34.499 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@4f147518)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:34.501 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7be4182a)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:34.501 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@7f39e23)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:34.503 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:34.503 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: -1, lastWrittenPosition: -1] 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:34.504 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:34.504 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:34.505 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:34.505 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:34.899 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Paused processing for partition 1 16:02:34.899 [] INFO io.camunda.zeebe.logstreams - Paused processing for stream stream-1 16:02:34.902 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:34.902 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:34.975 [] INFO io.camunda.zeebe.logstreams - Closed stream stream-1 16:02:35.163 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:35.163 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@82b217f)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6ecde0a, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@2d34a0f, configuration: Configuration(false)] 16:02:35.165 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:35.166 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@47d1d431)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@16a0356d, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@765cb53d, configuration: Configuration(false)] 16:02:35.167 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@6ec70acb)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@27fe2b7e, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@306279b1, configuration: Configuration(false)] 16:02:35.168 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@2304c7dd)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@75c30024, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@23b07039, configuration: Configuration(false)] 16:02:35.170 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@bc3a0f)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:35.171 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@5832fea)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:35.173 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@682ccd5e)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:35.174 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@483d49da)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:35.176 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:35.203 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: 19, lastWrittenPosition: 22] 16:02:35.203 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:35.203 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:35.204 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:35.204 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:35.663 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Paused processing for partition 1 16:02:35.663 [] INFO io.camunda.zeebe.logstreams - Paused processing for stream stream-1 16:02:35.666 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:35.667 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:35.711 [] INFO io.camunda.zeebe.logstreams - Closed stream stream-1 16:02:36.065 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Recovering state of partition 1 from snapshot 16:02:36.065 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7c6ce4c0)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@10bce2e0, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@4b428d95, configuration: Configuration(false)] 16:02:36.067 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.logstreams - Recovered state of partition 1 from snapshot at position -1 16:02:36.068 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@505b1058)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@4b0aa477, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@3c56f414, configuration: Configuration(false)] 16:02:36.069 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@63a7f82e)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@6b76edd2, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@75c1c5d7, configuration: Configuration(false)] 16:02:36.070 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@7ebc692f)), function-provider: io.camunda.zeebe.feel.impl.FeelFunctionProvider@2206985e, clock: io.camunda.zeebe.el.impl.ZeebeFeelEngineClock@7e42f9b1, configuration: Configuration(false)] 16:02:36.073 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@5d664d3)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:36.073 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@183e8057)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:36.076 [Broker-0-StreamProcessor-1] INFO org.camunda.dmn.DmnEngine - DMN-Engine created. [value-mapper: CompositeValueMapper(List(io.camunda.zeebe.feel.impl.MessagePackValueMapper@eec2f66)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, audit-loggers: List(), configuration: Configuration(false,false,false)] 16:02:36.076 [Broker-0-StreamProcessor-1] INFO org.camunda.feel.FeelEngine - Engine created. [value-mapper: CompositeValueMapper(List(org.camunda.dmn.NoUnpackValueMapper@670ad07b)), function-provider: org.camunda.feel.context.FunctionProvider$EmptyFunctionProvider$@60759723, clock: SystemClock, configuration: Configuration(false)] 16:02:36.079 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor starts replay of events. [snapshot-position: -1, replay-mode: PROCESSING] 16:02:36.106 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.processor - Processor finished replay, with [lastProcessedPosition: 37, lastWrittenPosition: 39] 16:02:36.106 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Starting processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:36.106 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Found 3 migration tasks: ProcessMessageSubscriptionSentTimeMigration, MessageSubscriptionSentTimeMigration, TemporaryVariableMigration 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping ProcessMessageSubscriptionSentTimeMigration migration (1/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping MessageSubscriptionSentTimeMigration migration (2/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Skipping TemporaryVariableMigration migration (3/3). It was determined it does not need to run right now. 16:02:36.107 [Broker-0-StreamProcessor-1] INFO io.camunda.zeebe.engine.state.migration - Completed processing of migration tasks (use LogLevel.DEBUG for more details) ... 16:02:36.107 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.engine.state.migration - Executed 0 migration tasks: 16:02:46.290 [] INFO io.camunda.zeebe.test - Test failed, following records were exported: 16:02:46.488 [] INFO io.camunda.zeebe.test - Compact log representation: -------- ['C'ommand/'E'event/'R'ejection] [valueType] [intent] - #[position]->#[source record position] P[partitionId]K[key] - [summary of value] P9K999 - key; #999 - record position; "ID" element/process id; @"elementid"/[P9K999] - element with ID and key Keys are decomposed into partition id and per partition key (e.g. 2251799813685253 -> P1K005). If single partition, the partition is omitted. Long IDs are shortened (e.g. 'startEvent_5d56488e-0570-416c-ba2d-36d2a3acea78' -> 'star..acea78' -------- C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C JOB_BATCH ACTIVATE - #23-> -1 -1 - "id_4_variables" max: 10 E JOB_BATCH ACTIVATED - #24->#23 K10 - "id_4_variables" 1/10 K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) C JOB COMPLETE - #25-> -1 K09 - K09 -1 retries, in <process ?[?]> with variables: {correlationKey=default_correlation_key} E JOB COMPLETED - #26->#25 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> with variables: {correlationKey=default_correlation_key} E PROC_EVNT TRIGGERING - #27->#25 K11 - @"id_4_va..es_task"[K08] in <process K01[K04]> with variables: {correlationKey=default_correlation_key} C PROC_INST COMPLETE - #28->#25 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #29->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E VAR CREATED - #30->#28 K12 - correlationKey->"default_correlation_key" in <process [K04]> E PROC_INST COMPLETED - #31->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #32->#28 K13 - SEQUENCE_FLOW "sequenc..ae31a49" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #33->#28 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #34->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_MSG_SUB CREATING - #35->#33 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #36->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> C MSG_SUB CREATE - #37-> -1 -1 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E MSG_SUB CREATED - #38->#37 K16 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) C PROC_MSG_SUB CREATE - #39-> -1 -1 - "message_id_0" (inter.) @[K14] in <process ?[K04]> (no vars) C DEPLOYMENT CREATE - #01-> -1 -1 - E PROC CREATED - #02->#01 K01 - process.xml -> "process_id_3" (version:1) E TIMER CREATED - #03->#01 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E DEPLOYMENT CREATED - #04->#01 K02 - process.xml E DEPLOYMENT FULLY_DISTR - #05->#01 K02 - C TIMER TRIGGER - #06-> -1 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E TIMER TRIGGERED - #07->#06 K03 - @"id_4"[ -1] in <process "K01"[?]> due T16:28:34.604 E PROC_EVNT TRIGGERING - #08->#06 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) C PROC_INST ACTIVATE - #09->#06 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #10->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #11->#09 K04 - PROCESS "process_id_3" in <process "process_id_3"[K04]> E PROC_EVNT TRIGGERED - #12->#09 K05 - @"id_4"[K01] in <process K01[K04]> (no vars) E PROC_INST ACTIVATING - #13->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST ACTIVATED - #14->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> C PROC_INST COMPLETE - #15->#09 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #16->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST COMPLETED - #17->#15 K06 - START_EVENT "id_4" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #18->#15 K07 - SEQUENCE_FLOW "sequenc..af851a0" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #19->#15 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #20->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E JOB CREATED - #21->#19 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #22->#19 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> C JOB_BATCH ACTIVATE - #23-> -1 -1 - "id_4_variables" max: 10 E JOB_BATCH ACTIVATED - #24->#23 K10 - "id_4_variables" 1/10 K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> (no vars) C JOB COMPLETE - #25-> -1 K09 - K09 -1 retries, in <process ?[?]> with variables: {correlationKey=default_correlation_key} E JOB COMPLETED - #26->#25 K09 - K09 "id_4_variables" @"id_4_va..es_task"[K08] 3 retries, in <process "process_id_3"[K04]> with variables: {correlationKey=default_correlation_key} E PROC_EVNT TRIGGERING - #27->#25 K11 - @"id_4_va..es_task"[K08] in <process K01[K04]> with variables: {correlationKey=default_correlation_key} C PROC_INST COMPLETE - #28->#25 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST COMPLETING - #29->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E VAR CREATED - #30->#28 K12 - correlationKey->"default_correlation_key" in <process [K04]> E PROC_INST COMPLETED - #31->#28 K08 - SERVICE_TASK "id_4_va..es_task" in <process "process_id_3"[K04]> E PROC_INST SEQ_FLOW_TAKEN - #32->#28 K13 - SEQUENCE_FLOW "sequenc..ae31a49" in <process "process_id_3"[K04]> C PROC_INST ACTIVATE - #33->#28 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_INST ACTIVATING - #34->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> E PROC_MSG_SUB CREATING - #35->#33 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E PROC_INST ACTIVATED - #36->#33 K14 - RECEIVE_TASK "id_0" in <process "process_id_3"[K04]> C MSG_SUB CREATE - #37-> -1 -1 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) E MSG_SUB CREATED - #38->#37 K16 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) C PROC_MSG_SUB CREATE - #39-> -1 -1 - "message_id_0" (inter.) @[K14] in <process ?[K04]> (no vars) E PROC_MSG_SUB CREATED - #40->#39 K15 - "message_id_0" (inter.) correlationKey: default_correlation_key @[K14] in <process "process_id_3"[K04]> (no vars) -------------- Deployed Processes ---------------------- process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> process.xml -> "process_id_3" (version:1)[K01] ------ <?xml version="1.0" encoding="UTF-8" standalone="no"?> <definitions xmlns:bpmndi="http://www.omg.org/spec/BPMN/20100524/DI" xmlns:dc="http://www.omg.org/spec/DD/20100524/DC" xmlns:di="http://www.omg.org/spec/DD/20100524/DI" xmlns:ns0="http://camunda.org/schema/zeebe/1.0" id="definitions_3ed4a392-8cbf-4963-9281-d27e5ebbddc8" targetNamespace="http://www.omg.org/spec/BPMN/20100524/MODEL" xmlns="http://www.omg.org/spec/BPMN/20100524/MODEL"> <process id="process_id_3" isExecutable="true"> <startEvent id="id_4" name="id_4"> <outgoing>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</outgoing> <timerEventDefinition id="timerEventDefinition_c050d071-da87-454a-ae61-b6e4fdb94d75"> <timeDate id="timeDate_caa9ac8c-f96a-4283-b596-e475352ed896">=now() + duration("PT26M")</timeDate> </timerEventDefinition> </startEvent> <serviceTask id="id_4_variables_task" name="id_4_variables_task"> <extensionElements> <ns0:taskDefinition type="id_4_variables"/> </extensionElements> <incoming>sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0</incoming> <outgoing>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</outgoing> </serviceTask> <sequenceFlow id="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" sourceRef="id_4" targetRef="id_4_variables_task"/> <receiveTask id="id_0" messageRef="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="id_0"> <incoming>sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49</incoming> <outgoing>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" sourceRef="id_4_variables_task" targetRef="id_0"/> <receiveTask id="id_1" messageRef="message_c5970535-a434-4638-a8f0-972bccba8a34" name="id_1"> <incoming>sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6</incoming> <outgoing>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</outgoing> </receiveTask> <sequenceFlow id="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" sourceRef="id_0" targetRef="id_1"/> <scriptTask id="id_2" name="id_2"> <extensionElements> <ns0:taskDefinition retries="3" type="job_id_2"/> </extensionElements> <incoming>sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e</incoming> <outgoing>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</outgoing> </scriptTask> <sequenceFlow id="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" sourceRef="id_1" targetRef="id_2"/> <endEvent id="id_5" name="id_5"> <incoming>sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081</incoming> </endEvent> <sequenceFlow id="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" sourceRef="id_2" targetRef="id_5"/> </process> <message id="message_a05f370f-6629-42ce-a4d5-79a6dbd5ca6a" name="message_id_0"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <message id="message_c5970535-a434-4638-a8f0-972bccba8a34" name="message_id_1"> <extensionElements> <ns0:subscription correlationKey="=correlationKey"/> </extensionElements> </message> <bpmndi:BPMNDiagram id="BPMNDiagram_f1e9a0b5-53a3-4983-96b6-d358ff7c21a7"> <bpmndi:BPMNPlane bpmnElement="process_id_3" id="BPMNPlane_7622abb7-fc72-479b-b9dc-976d09005b98"> <bpmndi:BPMNShape bpmnElement="id_4" id="BPMNShape_bdead3f2-07c4-481e-ba4b-9e16cc07de68"> <dc:Bounds height="36.0" width="36.0" x="100.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNShape bpmnElement="id_4_variables_task" id="BPMNShape_0f178e6e-9270-44a1-944c-9d6fac33c02b"> <dc:Bounds height="80.0" width="100.0" x="186.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_7cc2f8bd-02b0-4a4f-bf14-6729caf851a0" id="BPMNEdge_c422c337-88c4-4a0b-bdf5-b7a3f6004869"> <di:waypoint x="136.0" y="118.0"/> <di:waypoint x="186.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_0" id="BPMNShape_8153120f-4cb6-45cb-883f-7606d223ec24"> <dc:Bounds height="80.0" width="100.0" x="336.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_3976d7d2-d9bf-4778-b084-3ee95ae31a49" id="BPMNEdge_c7931cc9-8a1b-450e-932a-6d6566a6baa1"> <di:waypoint x="286.0" y="118.0"/> <di:waypoint x="336.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_1" id="BPMNShape_2d3cecd3-6aa7-47ad-ad24-e1ca95fa76aa"> <dc:Bounds height="80.0" width="100.0" x="486.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_4d283552-a181-45ab-b276-1c50a490f6d6" id="BPMNEdge_38e3fb27-5d0e-41d8-8b2a-96acedcf0043"> <di:waypoint x="436.0" y="118.0"/> <di:waypoint x="486.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_2" id="BPMNShape_355abee8-305d-4e8a-896d-7ebbd9916246"> <dc:Bounds height="80.0" width="100.0" x="636.0" y="78.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_d879c549-e345-4401-83c6-5452c5ef6b3e" id="BPMNEdge_d510c6fc-f962-4d1a-b392-2fcda5903c51"> <di:waypoint x="586.0" y="118.0"/> <di:waypoint x="636.0" y="118.0"/> </bpmndi:BPMNEdge> <bpmndi:BPMNShape bpmnElement="id_5" id="BPMNShape_c96c970d-27e3-49a8-b77d-2cae20596ec3"> <dc:Bounds height="36.0" width="36.0" x="786.0" y="100.0"/> </bpmndi:BPMNShape> <bpmndi:BPMNEdge bpmnElement="sequenceFlow_da486e04-76a3-414c-aa84-89cf3744d081" id="BPMNEdge_0a42e86c-cd8d-4ec8-8ca2-1d4cd9ccaec3"> <di:waypoint x="736.0" y="118.0"/> <di:waypoint x="786.0" y="118.0"/> </bpmndi:BPMNEdge> </bpmndi:BPMNPlane> </bpmndi:BPMNDiagram> </definitions> --------------- Decomposed keys (for debugging) ----------------- -1 <-> -1 K01 <-> 2251799813685249 K02 <-> 2251799813685250 K03 <-> 2251799813685251 K04 <-> 2251799813685252 K05 <-> 2251799813685253 K06 <-> 2251799813685254 K07 <-> 2251799813685255 K08 <-> 2251799813685256 K09 <-> 2251799813685257 K10 <-> 2251799813685258 K11 <-> 2251799813685259 K12 <-> 2251799813685260 K13 <-> 2251799813685261 K14 <-> 2251799813685262 K15 <-> 2251799813685263 K16 <-> 2251799813685264 16:02:46.495 [] DEBUG io.camunda.zeebe.util.buffer - Close stream processor 16:02:46.496 [Broker-0-StreamProcessor-1] DEBUG io.camunda.zeebe.logstreams - Closed stream processor controller Broker-0-StreamProcessor-1. 16:02:46.590 [Broker-0-LogStream-1] INFO io.camunda.zeebe.logstreams - Close appender for log stream stream-1 16:02:46.591 [stream-1-write-buffer] DEBUG io.camunda.zeebe.dispatcher - Dispatcher closed 16:02:46.591 [Broker-0-LogStream-1] INFO io.camunda.zeebe.logstreams - On closing logstream stream-1 close 6 readers 16:02:46.602 [] DEBUG io.camunda.zeebe.broker.test - Clean up test files on path /tmp/junit12212456625141852810 16:02:46.603 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers' 16:02:46.603 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-actors' 16:02:46.606 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-fs-workers': closed successfully 16:02:46.606 [] DEBUG io.camunda.zeebe.util.actor - Closing actor thread ground '-zb-actors': closed successfully 16:02:46.612 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Data of failed test case: TestDataRecord{processSeed=884380440262060521, executionPathSeed=2777154302604920046} 16:02:46.688 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Test case failed at: ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_4_variables_task,jobType=id_4_variables,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} 16:02:46.691 [] INFO io.camunda.zeebe.engine.processing.streamprocessor.FailedPropertyBasedTestDataPrinter - Execution path of failed test case: ScheduledExecutionStep{step=StepTriggerTimerStartEvent[timeToAdd=PT26M,variables={}], activationTime=PT0S, startTime=PT0S, endTime=PT26M, activationDuration=PT26M} --failed here--> ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_4_variables_task,jobType=id_4_variables,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepPublishMessage[correlationKeyValue=default_correlation_key,messageName=message_id_0,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepPublishMessage[correlationKeyValue=default_correlation_key,messageName=message_id_1,variables={correlationKey=default_correlation_key}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepActivateBPMNElement[elementId=id_2,variables={}], activationTime=PT26M, startTime=PT26M, endTime=PT26M, activationDuration=PT0S} ScheduledExecutionStep{step=StepActivateAndTimeoutJob[elementId=id_2,jobType=job_id_2,variables={}], activationTime=PT26M, startTime=PT26M, endTime=PT1H26M, activationDuration=PT1H} ScheduledExecutionStep{step=StepActivateAndCompleteJob[elementId=id_2,jobType=job_id_2,variables={}], activationTime=PT26M, startTime=PT1H26M, endTime=PT1H26M, activationDuration=PT1H} </pre> </details>
process
flaky replaystaterandomizedpropertytest shouldrestorestateateachstepinexecution summary failed in stable and stable how often does the test fail does it block your work no do we suspect that it is a real failure failures outline known failure cases e g a failed assertion and its stacktrace obtained from jenkins example assertion failure stacktrace org awaitility core conditiontimeoutexception condition with alias await that the replay state is equal to the processing state didn t complete within seconds because assertion condition defined as a io camunda zeebe engine processing streamprocessor replaystaterandomizedpropertytest multiple failures failure failure expecting map record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opened key to contain only record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opening key map entries not found record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opening key and map entries not expected record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opened key at replaystaterandomizedpropertytest lambda stopandrestartengineandcomparestates replaystaterandomizedpropertytest java at org awaitility core conditionawaiter await conditionawaiter java at org awaitility core assertioncondition await assertioncondition java at org awaitility core assertioncondition await assertioncondition java at org awaitility core conditionfactory until conditionfactory java at org awaitility core conditionfactory untilasserted conditionfactory java at io camunda zeebe engine processing streamprocessor replaystaterandomizedpropertytest stopandrestartengineandcomparestates replaystaterandomizedpropertytest java at io camunda zeebe engine processing streamprocessor replaystaterandomizedpropertytest shouldrestorestateateachstepinexecution replaystaterandomizedpropertytest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit rules testwatcher evaluate testwatcher java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit rules testwatcher evaluate testwatcher java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runner junitcore run junitcore java at org junit runner junitcore run junitcore java at org junit vintage engine execution runnerexecutor execute runnerexecutor java at org junit vintage engine vintagetestengine executeallchildren vintagetestengine java at org junit vintage engine vintagetestengine execute vintagetestengine java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator lambda execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator withinterceptedstreams engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlaunchersession delegatinglauncher execute defaultlaunchersession java at org junit platform launcher core sessionperrequestlauncher execute sessionperrequestlauncher java at org apache maven surefire junitplatform junitplatformprovider lambda execute junitplatformprovider java at java base java util iterator foreachremaining iterator java at org apache maven surefire junitplatform junitplatformprovider execute junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invokealltests junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invoke junitplatformprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by org assertj core error assertjmultiplefailureserror multiple failures failure failure expecting map record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opened key to contain only record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opening key map entries not found record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opening key and map entries not expected record subscriptionpartitionid processinstancekey elementinstancekey messagekey messagename message id variables ga interrupting true bpmnprocessid process id correlationkey default correlation key elementid id state state opened key at replaystaterandomizedpropertytest lambda stopandrestartengineandcomparestates replaystaterandomizedpropertytest java at jdk internal reflect newinstance unknown source at java base jdk internal reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java base java lang reflect constructor newinstancewithcaller constructor java at io camunda zeebe engine processing streamprocessor replaystaterandomizedpropertytest lambda stopandrestartengineandcomparestates replaystaterandomizedpropertytest java at org awaitility core assertioncondition lambda new assertioncondition java at org awaitility core conditionawaiter conditionpoller call conditionawaiter java at org awaitility core conditionawaiter conditionpoller call conditionawaiter java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java hypotheses logs logs standard output debug io camunda zeebe logstreams configured log appender back pressure at partition as appendervegascfg initiallimit maxconcurrency alphalimit betalimit window limiting is disabled debug io camunda zeebe logstreams recovering state of partition from snapshot info org camunda feel feelengine engine created info io camunda zeebe logstreams recovered state of partition from snapshot at position info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info io camunda zeebe processor processor starts replay of events info io camunda zeebe processor processor finished replay with info io camunda zeebe engine state migration starting processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration found migration tasks processmessagesubscriptionsenttimemigration messagesubscriptionsenttimemigration temporaryvariablemigration info io camunda zeebe engine state migration skipping processmessagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping messagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping temporaryvariablemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration completed processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration executed migration tasks debug io camunda zeebe logstreams paused processing for partition info io camunda zeebe logstreams paused processing for stream stream debug io camunda zeebe util buffer close stream processor debug io camunda zeebe logstreams closed stream processor controller broker streamprocessor info io camunda zeebe logstreams closed stream stream debug io camunda zeebe logstreams recovering state of partition from snapshot info org camunda feel feelengine engine created info io camunda zeebe logstreams recovered state of partition from snapshot at position info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info io camunda zeebe processor processor starts replay of events info io camunda zeebe processor processor finished replay with info io camunda zeebe engine state migration starting processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration found migration tasks processmessagesubscriptionsenttimemigration messagesubscriptionsenttimemigration temporaryvariablemigration info io camunda zeebe engine state migration skipping processmessagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping messagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping temporaryvariablemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration completed processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration executed migration tasks debug io camunda zeebe logstreams paused processing for partition info io camunda zeebe logstreams paused processing for stream stream debug io camunda zeebe util buffer close stream processor debug io camunda zeebe logstreams closed stream processor controller broker streamprocessor info io camunda zeebe logstreams closed stream stream debug io camunda zeebe logstreams recovering state of partition from snapshot info org camunda feel feelengine engine created info io camunda zeebe logstreams recovered state of partition from snapshot at position info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info org camunda dmn dmnengine dmn engine created info org camunda feel feelengine engine created info io camunda zeebe processor processor starts replay of events info io camunda zeebe processor processor finished replay with info io camunda zeebe engine state migration starting processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration found migration tasks processmessagesubscriptionsenttimemigration messagesubscriptionsenttimemigration temporaryvariablemigration info io camunda zeebe engine state migration skipping processmessagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping messagesubscriptionsenttimemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration skipping temporaryvariablemigration migration it was determined it does not need to run right now info io camunda zeebe engine state migration completed processing of migration tasks use loglevel debug for more details debug io camunda zeebe engine state migration executed migration tasks info io camunda zeebe test test failed following records were exported info io camunda zeebe test compact log representation p k key record position id element process id elementid element with id and key keys are decomposed into partition id and per partition key e g if single partition the partition is omitted long ids are shortened e g startevent star c deployment create e proc created process xml process id version e timer created id in due e deployment created process xml e deployment fully distr c timer trigger id in due e timer triggered id in due e proc evnt triggering id in no vars c proc inst activate process process id in e proc inst activating process process id in e proc inst activated process process id in e proc evnt triggered id in no vars e proc inst activating start event id in e proc inst activated start event id in c proc inst complete start event id in e proc inst completing start event id in e proc inst completed start event id in e proc inst seq flow taken sequence flow sequenc in c proc inst activate service task id va es task in e proc inst activating service task id va es task in e job created id variables id va es task retries in no vars e proc inst activated service task id va es task in c deployment create e proc created process xml process id version e timer created id in due e deployment created process xml e deployment fully distr c timer trigger id in due e timer triggered id in due e proc evnt triggering id in no vars c proc inst activate process process id in e proc inst activating process process id in e proc inst activated process process id in e proc evnt triggered id in no vars e proc inst activating start event id in e proc inst activated start event id in c proc inst complete start event id in e proc inst completing start event id in e proc inst completed start event id in e proc inst seq flow taken sequence flow sequenc in c proc inst activate service task id va es task in e proc inst activating service task id va es task in e job created id variables id va es task retries in no vars e proc inst activated service task id va es task in c job batch activate id variables max e job batch activated id variables id variables id va es task retries in no vars c job complete retries in with variables correlationkey default correlation key e job completed id variables id va es task retries in with variables correlationkey default correlation key e proc evnt triggering id va es task in with variables correlationkey default correlation key c proc inst complete service task id va es task in e proc inst completing service task id va es task in e var created correlationkey default correlation key in e proc inst completed service task id va es task in e proc inst seq flow taken sequence flow sequenc in c proc inst activate receive task id in e proc inst activating receive task id in e proc msg sub creating message id inter correlationkey default correlation key in no vars e proc inst activated receive task id in c msg sub create message id inter correlationkey default correlation key in no vars e msg sub created message id inter correlationkey default correlation key in no vars c proc msg sub create message id inter in no vars c deployment create e proc created process xml process id version e timer created id in due e deployment created process xml e deployment fully distr c timer trigger id in due e timer triggered id in due e proc evnt triggering id in no vars c proc inst activate process process id in e proc inst activating process process id in e proc inst activated process process id in e proc evnt triggered id in no vars e proc inst activating start event id in e proc inst activated start event id in c proc inst complete start event id in e proc inst completing start event id in e proc inst completed start event id in e proc inst seq flow taken sequence flow sequenc in c proc inst activate service task id va es task in e proc inst activating service task id va es task in e job created id variables id va es task retries in no vars e proc inst activated service task id va es task in c job batch activate id variables max e job batch activated id variables id variables id va es task retries in no vars c job complete retries in with variables correlationkey default correlation key e job completed id variables id va es task retries in with variables correlationkey default correlation key e proc evnt triggering id va es task in with variables correlationkey default correlation key c proc inst complete service task id va es task in e proc inst completing service task id va es task in e var created correlationkey default correlation key in e proc inst completed service task id va es task in e proc inst seq flow taken sequence flow sequenc in c proc inst activate receive task id in e proc inst activating receive task id in e proc msg sub creating message id inter correlationkey default correlation key in no vars e proc inst activated receive task id in c msg sub create message id inter correlationkey default correlation key in no vars e msg sub created message id inter correlationkey default correlation key in no vars c proc msg sub create message id inter in no vars e proc msg sub created message id inter correlationkey default correlation key in no vars deployed processes process xml process id version definitions xmlns bpmndi xmlns dc xmlns di xmlns id definitions targetnamespace xmlns sequenceflow now duration sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow process xml process id version definitions xmlns bpmndi xmlns dc xmlns di xmlns id definitions targetnamespace xmlns sequenceflow now duration sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow process xml process id version definitions xmlns bpmndi xmlns dc xmlns di xmlns id definitions targetnamespace xmlns sequenceflow now duration sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow sequenceflow decomposed keys for debugging debug io camunda zeebe util buffer close stream processor debug io camunda zeebe logstreams closed stream processor controller broker streamprocessor info io camunda zeebe logstreams close appender for log stream stream debug io camunda zeebe dispatcher dispatcher closed info io camunda zeebe logstreams on closing logstream stream close readers debug io camunda zeebe broker test clean up test files on path tmp debug io camunda zeebe util actor closing actor thread ground zb fs workers debug io camunda zeebe util actor closing actor thread ground zb actors debug io camunda zeebe util actor closing actor thread ground zb fs workers closed successfully debug io camunda zeebe util actor closing actor thread ground zb actors closed successfully info io camunda zeebe engine processing streamprocessor failedpropertybasedtestdataprinter data of failed test case testdatarecord processseed executionpathseed info io camunda zeebe engine processing streamprocessor failedpropertybasedtestdataprinter test case failed at scheduledexecutionstep step stepactivateandcompletejob activationtime starttime endtime activationduration info io camunda zeebe engine processing streamprocessor failedpropertybasedtestdataprinter execution path of failed test case scheduledexecutionstep step steptriggertimerstartevent activationtime starttime endtime activationduration failed here scheduledexecutionstep step stepactivateandcompletejob activationtime starttime endtime activationduration scheduledexecutionstep step steppublishmessage activationtime starttime endtime activationduration scheduledexecutionstep step steppublishmessage activationtime starttime endtime activationduration scheduledexecutionstep step stepactivatebpmnelement activationtime starttime endtime activationduration scheduledexecutionstep step stepactivateandtimeoutjob activationtime starttime endtime activationduration scheduledexecutionstep step stepactivateandcompletejob activationtime starttime endtime activationduration
1
1,079
3,541,524,295
IssuesEvent
2016-01-19 01:42:58
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
Реализовать создание двух отдельных сообщений в сущности SubjectMessage при вызове сервиса /setMessageFeedback_Indirectly
active In process of testing test _wf-central
подобно тому, как это реализовано в сервисе /setMessageRate 1) для рейтинга тип сообщения должен быть 2 Хеадер(sHead): Уточнююча оцінка о відпрацованій послузі по заяві " + sID_Order Тело(sBody): Оцінка sID_Rate_Indirectly (по шкалі від 2 до 5) 2) для отзыва тип сообщения должен быть 1 Хеадер(sHead): Уточнюючий відгук о відпрацованій послузі по заяві " + sID_Order Тело(sBody): sBody_Indirectly
1.0
Реализовать создание двух отдельных сообщений в сущности SubjectMessage при вызове сервиса /setMessageFeedback_Indirectly - подобно тому, как это реализовано в сервисе /setMessageRate 1) для рейтинга тип сообщения должен быть 2 Хеадер(sHead): Уточнююча оцінка о відпрацованій послузі по заяві " + sID_Order Тело(sBody): Оцінка sID_Rate_Indirectly (по шкалі від 2 до 5) 2) для отзыва тип сообщения должен быть 1 Хеадер(sHead): Уточнюючий відгук о відпрацованій послузі по заяві " + sID_Order Тело(sBody): sBody_Indirectly
process
реализовать создание двух отдельных сообщений в сущности subjectmessage при вызове сервиса setmessagefeedback indirectly подобно тому как это реализовано в сервисе setmessagerate для рейтинга тип сообщения должен быть хеадер shead уточнююча оцінка о відпрацованій послузі по заяві sid order тело sbody оцінка sid rate indirectly по шкалі від до для отзыва тип сообщения должен быть хеадер shead уточнюючий відгук о відпрацованій послузі по заяві sid order тело sbody sbody indirectly
1
5,119
7,889,212,611
IssuesEvent
2018-06-28 02:39:31
raxod502/straight.el
https://api.github.com/repos/raxod502/straight.el
closed
Minor gripe: *straight-process* buffer contains peaceful failures
discussion process-buffer ux
Just did a [big update](https://github.com/tummychow/emacs/commit/368d5cf089fc85c0ebc1e70377152c7231b872c5) and now every time I boot up emacs I see this in `*straight-process*`: ``` $ cd /home/sjung/.emacs.d/straight/repos/epkgs/ $ git config --file .gitmodules --get submodule.let-alist.url [Return code: 1] $ cd /home/sjung/.emacs.d/straight/repos/epkgs/ $ git config --file .gitmodules --get submodule.let-alist.url [Return code: 1] ``` spent a while debugging it before realizing that this really isn't even an error (`let-alist` is a built-in). This has probably always had a nonzero return code, but wasn't opening a buffer for it until now. I would prefer to only see this buffer spawned for genuine breakages that need my attention.
1.0
Minor gripe: *straight-process* buffer contains peaceful failures - Just did a [big update](https://github.com/tummychow/emacs/commit/368d5cf089fc85c0ebc1e70377152c7231b872c5) and now every time I boot up emacs I see this in `*straight-process*`: ``` $ cd /home/sjung/.emacs.d/straight/repos/epkgs/ $ git config --file .gitmodules --get submodule.let-alist.url [Return code: 1] $ cd /home/sjung/.emacs.d/straight/repos/epkgs/ $ git config --file .gitmodules --get submodule.let-alist.url [Return code: 1] ``` spent a while debugging it before realizing that this really isn't even an error (`let-alist` is a built-in). This has probably always had a nonzero return code, but wasn't opening a buffer for it until now. I would prefer to only see this buffer spawned for genuine breakages that need my attention.
process
minor gripe straight process buffer contains peaceful failures just did a and now every time i boot up emacs i see this in straight process cd home sjung emacs d straight repos epkgs git config file gitmodules get submodule let alist url cd home sjung emacs d straight repos epkgs git config file gitmodules get submodule let alist url spent a while debugging it before realizing that this really isn t even an error let alist is a built in this has probably always had a nonzero return code but wasn t opening a buffer for it until now i would prefer to only see this buffer spawned for genuine breakages that need my attention
1
12,159
14,741,492,852
IssuesEvent
2021-01-07 10:42:19
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
error in SA Billing - taxes selection dropping off when editing an account
anc-process anp-0.5 ant-bug ant-parent/primary ant-support
In GitLab by @kdjstudios on Jan 22, 2019, 15:08 **Submitted by:** From: Sharon Carver <scarver@laseranswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-22-15253 **Server:** External (Both?) **Client/Site:** Laser (All?) **Account:** All **Issue:** Sharon Wrote: > Yes, we are still experiencing issues. Now it is with taxes. When I edit an existing account, the taxes are no longer selected. Please see account 7959 where taxes were not charged in Dec or Jan as a result. I did not notice this until the customer was past due and I was trying to charge a credit card for the first time and I noticed no taxes were on their invoice. > > Just a moment ago, I edited customer 2194 and the taxes were not selected. I wrote: > Thank you for notifying us. I just want to confirm that no one on your team had removed the taxes from those accounts? We have informed the support team and will follow up with their research and response. Sharon Wrote: > No, no one here has removed the taxes from those accounts. > > It looks like when we go into edit an account, the taxes selection is dropping off. Not from our doing. I just tested another account that gets charged tax, 2121 and it happened. However, after I select taxes and save, it then is 'sticking' because taxes are now attached. I tried it with 2121, 2194, 5158, 7823. > > I suggest you do the same with another account to see if it happens to you too. Just click on edit account, scroll down to the taxes box and you will see nothing is selected - try 2118 or 2131. I think this will happen with any taxable account you choose. Something is wrong with the edit function of SA billing. I wrote: > Thank you for the confirmation and additional information. I was able to validate and reproduce the issue where the taxes are not being displayed on the edit account screen. We have the support team looking into this and will keep you updated.
1.0
error in SA Billing - taxes selection dropping off when editing an account - In GitLab by @kdjstudios on Jan 22, 2019, 15:08 **Submitted by:** From: Sharon Carver <scarver@laseranswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-22-15253 **Server:** External (Both?) **Client/Site:** Laser (All?) **Account:** All **Issue:** Sharon Wrote: > Yes, we are still experiencing issues. Now it is with taxes. When I edit an existing account, the taxes are no longer selected. Please see account 7959 where taxes were not charged in Dec or Jan as a result. I did not notice this until the customer was past due and I was trying to charge a credit card for the first time and I noticed no taxes were on their invoice. > > Just a moment ago, I edited customer 2194 and the taxes were not selected. I wrote: > Thank you for notifying us. I just want to confirm that no one on your team had removed the taxes from those accounts? We have informed the support team and will follow up with their research and response. Sharon Wrote: > No, no one here has removed the taxes from those accounts. > > It looks like when we go into edit an account, the taxes selection is dropping off. Not from our doing. I just tested another account that gets charged tax, 2121 and it happened. However, after I select taxes and save, it then is 'sticking' because taxes are now attached. I tried it with 2121, 2194, 5158, 7823. > > I suggest you do the same with another account to see if it happens to you too. Just click on edit account, scroll down to the taxes box and you will see nothing is selected - try 2118 or 2131. I think this will happen with any taxable account you choose. Something is wrong with the edit function of SA billing. I wrote: > Thank you for the confirmation and additional information. I was able to validate and reproduce the issue where the taxes are not being displayed on the edit account screen. We have the support team looking into this and will keep you updated.
process
error in sa billing taxes selection dropping off when editing an account in gitlab by kdjstudios on jan submitted by from sharon carver helpdesk server external both client site laser all account all issue sharon wrote yes we are still experiencing issues now it is with taxes when i edit an existing account the taxes are no longer selected please see account where taxes were not charged in dec or jan as a result i did not notice this until the customer was past due and i was trying to charge a credit card for the first time and i noticed no taxes were on their invoice just a moment ago i edited customer and the taxes were not selected i wrote thank you for notifying us i just want to confirm that no one on your team had removed the taxes from those accounts we have informed the support team and will follow up with their research and response sharon wrote no no one here has removed the taxes from those accounts it looks like when we go into edit an account the taxes selection is dropping off not from our doing i just tested another account that gets charged tax and it happened however after i select taxes and save it then is sticking because taxes are now attached i tried it with i suggest you do the same with another account to see if it happens to you too just click on edit account scroll down to the taxes box and you will see nothing is selected try or i think this will happen with any taxable account you choose something is wrong with the edit function of sa billing i wrote thank you for the confirmation and additional information i was able to validate and reproduce the issue where the taxes are not being displayed on the edit account screen we have the support team looking into this and will keep you updated
1
201,897
23,045,943,095
IssuesEvent
2022-07-23 22:39:53
turkdevops/javascript-sdk
https://api.github.com/repos/turkdevops/javascript-sdk
closed
CVE-2020-11022 (Medium) detected in jquery-1.7.2.min.js, jquery-1.8.1.min.js - autoclosed
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.8.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/dmd/node_modules/marked/www/demo.html</p> <p>Path to vulnerable library: /node_modules/dmd/node_modules/marked/www/demo.html,/node_modules/jmespath/index.html,/node_modules/jsdoc/node_modules/marked/www/demo.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-1.7.2.min.js, jquery-1.8.1.min.js - autoclosed - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.7.2.min.js</b>, <b>jquery-1.8.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/dmd/node_modules/marked/www/demo.html</p> <p>Path to vulnerable library: /node_modules/dmd/node_modules/marked/www/demo.html,/node_modules/jmespath/index.html,/node_modules/jsdoc/node_modules/marked/www/demo.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js jquery min js autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file node modules dmd node modules marked www demo html path to vulnerable library node modules dmd node modules marked www demo html node modules jmespath index html node modules jsdoc node modules marked www demo html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node modules redeyed examples browser index html path to vulnerable library node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
0
9,637
12,601,174,317
IssuesEvent
2020-06-11 09:24:11
prisma/e2e-tests
https://api.github.com/repos/prisma/e2e-tests
closed
Add zeit/pkg test
kind/feature process/candidate team/other
We should add a test which compiles the Prisma CLI with zeit/pkg and then runs a few commands against the resulting binary. We need zeit/pkg for packaging and delivering the ClI with generators such as the [go client](https://github.com/prisma/prisma-client-go).
1.0
Add zeit/pkg test - We should add a test which compiles the Prisma CLI with zeit/pkg and then runs a few commands against the resulting binary. We need zeit/pkg for packaging and delivering the ClI with generators such as the [go client](https://github.com/prisma/prisma-client-go).
process
add zeit pkg test we should add a test which compiles the prisma cli with zeit pkg and then runs a few commands against the resulting binary we need zeit pkg for packaging and delivering the cli with generators such as the
1
7,681
10,762,610,256
IssuesEvent
2019-11-01 00:20:01
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Question: quorum sensing GO:0009372
Other term-related request multi-species process
Quoring sensing is defined: The process in which single-celled organisms monitor their population density by detecting the concentration of small, diffusible signal molecules produced by the cells themselves. PMID:15716452 PMID:8288518 So is quorum sension only the detection and signalling? Or is it also the production of the molecules that are sensed by the "quorum sensing" These would be quite different in scope. We are not sure whether production of the signalling molecules is caussally upstream, regulatory, or part of the process
1.0
Question: quorum sensing GO:0009372 - Quoring sensing is defined: The process in which single-celled organisms monitor their population density by detecting the concentration of small, diffusible signal molecules produced by the cells themselves. PMID:15716452 PMID:8288518 So is quorum sension only the detection and signalling? Or is it also the production of the molecules that are sensed by the "quorum sensing" These would be quite different in scope. We are not sure whether production of the signalling molecules is caussally upstream, regulatory, or part of the process
process
question quorum sensing go quoring sensing is defined the process in which single celled organisms monitor their population density by detecting the concentration of small diffusible signal molecules produced by the cells themselves pmid pmid so is quorum sension only the detection and signalling or is it also the production of the molecules that are sensed by the quorum sensing these would be quite different in scope we are not sure whether production of the signalling molecules is caussally upstream regulatory or part of the process
1
75,344
25,780,699,097
IssuesEvent
2022-12-09 15:41:08
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
CMS — 508-defect-2 [AXE-CORE]: Heading levels SHOULD increase by one - alert headings
⭐️ Public Websites 508/Accessibility cms vsa 508-defect-2 508-issue-headings
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) ```diff ! Team affected: VSA-public-websites; Project found in: AMA Follow-on ``` Also documented in https://github.com/department-of-veterans-affairs/va.gov-team/issues/6375. **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Headings on a web page **should** progress, increasing by one, in sections of content. There are a number of headings throughout the pages that don't adhere to the heading order of H1, H2, H3, H4, etc. 1. Decision reviews and appeals — https://preview-prod.vfs.va.gov_preview_nodeid=3071 | Section:nth-child(1) > h4 — `<h4>Message us</h4>` 1. Supplemental Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3011 | `<h3 id="can-i-file-a-supplemental-clai">` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Board Appeals — https://preview-prod.vfs.va.gov/preview?nodeId=3029 | `<h3 id="can-i-request-a-board-appeal">Can I request a Board Appeal?</h3>` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Contested Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3030 | `<h3 class="use-alert-heading" id="mark-your-calendar">` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Fiduciary claims — https://preview-prod.vfs.va.gov/preview?nodeId=3016 | `<h3 class="use-alert-heading" id="mark-your-calendar">` 1. Insurance claims — https://preview-prod.vfs.va.gov/preview?nodeId=3025 | `<h3 class="use-alert-heading" id="mark-your-calendar">` 1. Get help with review request — https://preview-prod.vfs.va.gov/preview?nodeId=3009 | `<h3 itemprop="name" id="what-does-it-take-to-be-an-acc">` 1. Board hearings with a VLJ — https://preview-prod.vfs.va.gov/preview?nodeId=3010 | `<h3 class="use-alert-heading id="the-process-for-board-hearings">` 1. [added 06/09/20, during Full A11y Audit] [Higher-Level Reviews](https://www.va.gov/decision-reviews/higher-level-review/) `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. [added 06/09/20, during Full A11y Audit][FAQ](https://www.va.gov/decision-reviews/faq/) `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` ### Why it Matters The underlying purpose of headers is to convey the structure of the page. For sighted users, the same purpose is achieved using different sizes of text. Text size, however, is not helpful for users of screen readers, because a screen reader identifies a header only if it is properly marked-up. When heading elements are applied correctly, the page becomes much easier to navigate for screen reader users and sighted users alike. In the same way that sighted users can glance at a page and get a sense of its contents, users of screen readers can do the same by navigating through headings. Well written and properly ordered headings can save users, especially those who use screen readers, a lot of time and frustration. The purpose of headings is to describe the structure of the webpage, not just highlight important text. They should be brief, clear, unique, and marked with h1 through h6 elements applied in hierarchical order. All of these qualities make headings valuable tools for screen reader users. Similar to the way sighted users can glance at a page and get a sense of its contents, screen reader users can navigate through headings. Well written and properly ordered headings can save screen reader time and frustration. In addition to making the page more accessible, headings have other benefits since search engines use headings when filtering, ordering, and displaying results. Improving the accessibility of your site can also have the effect of making your page more findable. ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to navigate the hierarchy of the page content using heading levels to save time and frustration. ## Environment * Operating System: all * Browser: all * Screenreading device: any * Server destination: staging & production ## Steps to Recreate 1. Enter any of the above urls in browser, in the VA environment 1. Have developer tools open, and the axe browser extension loaded 1. Enter Urgent care for facility type and Community urgent care providers for service type 1. Run an axe audit 1. Verify that there is an error of "Heading levels should only increase by one" ## Possible Fixes (optional) The recommendation is the h3 become an h2 with the h3 utility class. ## WCAG or Vendor Guidance (optional) * [axe-core 3.4 - Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/3.4/heading-order) * [MDN dialog element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dialog) * [MDN ARIA: alert role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/Alert_Role) * [MDN Description List element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dl)
1.0
CMS — 508-defect-2 [AXE-CORE]: Heading levels SHOULD increase by one - alert headings - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) ```diff ! Team affected: VSA-public-websites; Project found in: AMA Follow-on ``` Also documented in https://github.com/department-of-veterans-affairs/va.gov-team/issues/6375. **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Headings on a web page **should** progress, increasing by one, in sections of content. There are a number of headings throughout the pages that don't adhere to the heading order of H1, H2, H3, H4, etc. 1. Decision reviews and appeals — https://preview-prod.vfs.va.gov_preview_nodeid=3071 | Section:nth-child(1) > h4 — `<h4>Message us</h4>` 1. Supplemental Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3011 | `<h3 id="can-i-file-a-supplemental-clai">` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Board Appeals — https://preview-prod.vfs.va.gov/preview?nodeId=3029 | `<h3 id="can-i-request-a-board-appeal">Can I request a Board Appeal?</h3>` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Contested Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3030 | `<h3 class="use-alert-heading" id="mark-your-calendar">` and `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. Fiduciary claims — https://preview-prod.vfs.va.gov/preview?nodeId=3016 | `<h3 class="use-alert-heading" id="mark-your-calendar">` 1. Insurance claims — https://preview-prod.vfs.va.gov/preview?nodeId=3025 | `<h3 class="use-alert-heading" id="mark-your-calendar">` 1. Get help with review request — https://preview-prod.vfs.va.gov/preview?nodeId=3009 | `<h3 itemprop="name" id="what-does-it-take-to-be-an-acc">` 1. Board hearings with a VLJ — https://preview-prod.vfs.va.gov/preview?nodeId=3010 | `<h3 class="use-alert-heading id="the-process-for-board-hearings">` 1. [added 06/09/20, during Full A11y Audit] [Higher-Level Reviews](https://www.va.gov/decision-reviews/higher-level-review/) `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` 1. [added 06/09/20, during Full A11y Audit][FAQ](https://www.va.gov/decision-reviews/faq/) `<h3 class="usa-alert-heading" id="you-can-still-file-a-claim-and">` ### Why it Matters The underlying purpose of headers is to convey the structure of the page. For sighted users, the same purpose is achieved using different sizes of text. Text size, however, is not helpful for users of screen readers, because a screen reader identifies a header only if it is properly marked-up. When heading elements are applied correctly, the page becomes much easier to navigate for screen reader users and sighted users alike. In the same way that sighted users can glance at a page and get a sense of its contents, users of screen readers can do the same by navigating through headings. Well written and properly ordered headings can save users, especially those who use screen readers, a lot of time and frustration. The purpose of headings is to describe the structure of the webpage, not just highlight important text. They should be brief, clear, unique, and marked with h1 through h6 elements applied in hierarchical order. All of these qualities make headings valuable tools for screen reader users. Similar to the way sighted users can glance at a page and get a sense of its contents, screen reader users can navigate through headings. Well written and properly ordered headings can save screen reader time and frustration. In addition to making the page more accessible, headings have other benefits since search engines use headings when filtering, ordering, and displaying results. Improving the accessibility of your site can also have the effect of making your page more findable. ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to navigate the hierarchy of the page content using heading levels to save time and frustration. ## Environment * Operating System: all * Browser: all * Screenreading device: any * Server destination: staging & production ## Steps to Recreate 1. Enter any of the above urls in browser, in the VA environment 1. Have developer tools open, and the axe browser extension loaded 1. Enter Urgent care for facility type and Community urgent care providers for service type 1. Run an axe audit 1. Verify that there is an error of "Heading levels should only increase by one" ## Possible Fixes (optional) The recommendation is the h3 become an h2 with the h3 utility class. ## WCAG or Vendor Guidance (optional) * [axe-core 3.4 - Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/3.4/heading-order) * [MDN dialog element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dialog) * [MDN ARIA: alert role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/Alert_Role) * [MDN Description List element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dl)
non_process
cms —  defect heading levels should increase by one alert headings diff team affected vsa public websites project found in ama follow on also documented in feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description headings on a web page should progress increasing by one in sections of content there are a number of headings throughout the pages that don t adhere to the heading order of etc decision reviews and appeals — section nth child — message us supplemental claims — and board appeals — can i request a board appeal and contested claims — and fiduciary claims — insurance claims — get help with review request — board hearings with a vlj — why it matters the underlying purpose of headers is to convey the structure of the page for sighted users the same purpose is achieved using different sizes of text text size however is not helpful for users of screen readers because a screen reader identifies a header only if it is properly marked up when heading elements are applied correctly the page becomes much easier to navigate for screen reader users and sighted users alike in the same way that sighted users can glance at a page and get a sense of its contents users of screen readers can do the same by navigating through headings well written and properly ordered headings can save users especially those who use screen readers a lot of time and frustration the purpose of headings is to describe the structure of the webpage not just highlight important text they should be brief clear unique and marked with through elements applied in hierarchical order all of these qualities make headings valuable tools for screen reader users similar to the way sighted users can glance at a page and get a sense of its contents screen reader users can navigate through headings well written and properly ordered headings can save screen reader time and frustration in addition to making the page more accessible headings have other benefits since search engines use headings when filtering ordering and displaying results improving the accessibility of your site can also have the effect of making your page more findable point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to navigate the hierarchy of the page content using heading levels to save time and frustration environment operating system all browser all screenreading device any server destination staging production steps to recreate enter any of the above urls in browser in the va environment have developer tools open and the axe browser extension loaded enter urgent care for facility type and community urgent care providers for service type run an axe audit verify that there is an error of heading levels should only increase by one possible fixes optional the recommendation is the become an with the utility class wcag or vendor guidance optional
0
3,478
6,553,010,953
IssuesEvent
2017-09-05 20:39:41
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
Generic GeoJSON importer
processed story
Should be a basic way and tutorial to import GeoJSON POI's. Perhaps can be done as part of pelias/pelias#117.
1.0
Generic GeoJSON importer - Should be a basic way and tutorial to import GeoJSON POI's. Perhaps can be done as part of pelias/pelias#117.
process
generic geojson importer should be a basic way and tutorial to import geojson poi s perhaps can be done as part of pelias pelias
1
48,981
13,185,181,504
IssuesEvent
2020-08-12 20:53:06
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
geometry renderer specifies center of detector (Trac #558)
Incomplete Migration Migrated from Trac defect glshovel
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/558 , reported by troy and owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-07-20T22:14:29", "description": "for km3net... (0,0,0) isn't in the center of their detector.", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1248128069000000", "component": "glshovel", "summary": "geometry renderer specifies center of detector", "priority": "normal", "keywords": "", "time": "2009-06-12T13:52:20", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
geometry renderer specifies center of detector (Trac #558) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/558 , reported by troy and owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-07-20T22:14:29", "description": "for km3net... (0,0,0) isn't in the center of their detector.", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1248128069000000", "component": "glshovel", "summary": "geometry renderer specifies center of detector", "priority": "normal", "keywords": "", "time": "2009-06-12T13:52:20", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
non_process
geometry renderer specifies center of detector trac migrated from reported by troy and owned by troy json status closed changetime description for isn t in the center of their detector reporter troy cc resolution fixed ts component glshovel summary geometry renderer specifies center of detector priority normal keywords time milestone owner troy type defect
0
283,589
30,913,492,189
IssuesEvent
2023-08-05 02:03:59
hshivhare67/kernel_v4.19.72
https://api.github.com/repos/hshivhare67/kernel_v4.19.72
reopened
CVE-2023-0045 (High) detected in linuxlinux-4.19.282
Mend: dependency security vulnerability
## CVE-2023-0045 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The current implementation of the prctl syscall does not issue an IBPB immediately during the syscall. The ib_prctl_set  function updates the Thread Information Flags (TIFs) for the task and updates the SPEC_CTRL MSR on the function __speculation_ctrl_update, but the IBPB is only issued on the next schedule, when the TIF bits are checked. This leaves the victim vulnerable to values already injected on the BTB, prior to the prctl syscall.  The patch that added the support for the conditional mitigation via prctl (ib_prctl_set) dates back to the kernel 4.9.176. We recommend upgrading past commit a664ec9158eeddd75121d39c9a0758016097fa96 <p>Publish Date: 2023-04-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0045>CVE-2023-0045</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0045">https://www.linuxkernelcves.com/cves/CVE-2023-0045</a></p> <p>Release Date: 2023-01-04</p> <p>Fix Resolution: v5.4.229,v5.10.163,v5.15.87,v6.0.19,v6.1.5,v6.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-0045 (High) detected in linuxlinux-4.19.282 - ## CVE-2023-0045 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The current implementation of the prctl syscall does not issue an IBPB immediately during the syscall. The ib_prctl_set  function updates the Thread Information Flags (TIFs) for the task and updates the SPEC_CTRL MSR on the function __speculation_ctrl_update, but the IBPB is only issued on the next schedule, when the TIF bits are checked. This leaves the victim vulnerable to values already injected on the BTB, prior to the prctl syscall.  The patch that added the support for the conditional mitigation via prctl (ib_prctl_set) dates back to the kernel 4.9.176. We recommend upgrading past commit a664ec9158eeddd75121d39c9a0758016097fa96 <p>Publish Date: 2023-04-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0045>CVE-2023-0045</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0045">https://www.linuxkernelcves.com/cves/CVE-2023-0045</a></p> <p>Release Date: 2023-01-04</p> <p>Fix Resolution: v5.4.229,v5.10.163,v5.15.87,v6.0.19,v6.1.5,v6.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the current implementation of the prctl syscall does not issue an ibpb immediately during the syscall the ib prctl set  function updates the thread information flags tifs for the task and updates the spec ctrl msr on the function speculation ctrl update but the ibpb is only issued on the next schedule when the tif bits are checked this leaves the victim vulnerable to values already injected on the btb prior to the prctl syscall  the patch that added the support for the conditional mitigation via prctl ib prctl set dates back to the kernel we recommend upgrading past commit  publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
19,686
26,036,218,572
IssuesEvent
2022-12-22 05:24:24
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Start stop v1 removed from marketplace
automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri1
Hello, The note at the top of the page mentions that version 1 will soon be removed from the marketplace but it is already removed. Could you please update the note? _> Note > > Start/Stop VM during off-hours, **version 1 is currently being deprecated and will be unavailable from the marketplace soon.** We recommend that you start using [version 2](https://learn.microsoft.com/en-us/azure/azure-functions/start-stop-vms/overview), which is now generally available. The new version offers all existing capabilities and provides new features, such as multi-subscription support from a single Start/Stop instance. If you have the version 1 solution already deployed, you can still use the feature, and we will provide support until further announcement._ --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://learn.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
1.0
Start stop v1 removed from marketplace - Hello, The note at the top of the page mentions that version 1 will soon be removed from the marketplace but it is already removed. Could you please update the note? _> Note > > Start/Stop VM during off-hours, **version 1 is currently being deprecated and will be unavailable from the marketplace soon.** We recommend that you start using [version 2](https://learn.microsoft.com/en-us/azure/azure-functions/start-stop-vms/overview), which is now generally available. The new version offers all existing capabilities and provides new features, such as multi-subscription support from a single Start/Stop instance. If you have the version 1 solution already deployed, you can still use the feature, and we will provide support until further announcement._ --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://learn.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
process
start stop removed from marketplace hello the note at the top of the page mentions that version will soon be removed from the marketplace but it is already removed could you please update the note note start stop vm during off hours version is currently being deprecated and will be unavailable from the marketplace soon we recommend that you start using which is now generally available the new version offers all existing capabilities and provides new features such as multi subscription support from a single start stop instance if you have the version solution already deployed you can still use the feature and we will provide support until further announcement document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha
1
15,288
19,293,376,245
IssuesEvent
2021-12-12 06:43:05
redwoodjs/redwood
https://api.github.com/repos/redwoodjs/redwood
reopened
Add a `scripts/tsconfig.json` to the create-redwood-app template
triage/processing
Hello, Making this as an issue in the off-chance there are plans to support something similar in a different way. I have no issue pushing this configuration (or another) up myself. --- I went to update the `scripts/seed.ts` file that gets generated with a Redwood project and was greeted with a TypeScript error: <details> <summary>TypeScript resolution error</summary> ![image](https://user-images.githubusercontent.com/25166787/142083693-dfd29c6c-b4d1-4353-9cab-731c949bd912.png) </details> The fix is easy enough, here is the `tsconfig.json` file I added to the `scripts/` directory. <details> <summary>scripts/tsconfig.json</summary> ```YAML { "compilerOptions": { "noEmit": true, "allowJs": true, "esModuleInterop": true, "target": "esnext", "module": "esnext", "moduleResolution": "node", "resolveJsonModule": true, "baseUrl": "./", "paths": { "api/*": ["../api/*"], "web/*": ["../web/*"] }, "typeRoots": ["../node_modules/@types", "./node_modules/@types"], "types": ["jest"], }, "include": [ "./", "../.redwood/types/includes/all-*", ] } ``` </details> Was having the `$` character before the resolution something that was desired/necessary? It felt more natural and consistent with each side to leave it off.
1.0
Add a `scripts/tsconfig.json` to the create-redwood-app template - Hello, Making this as an issue in the off-chance there are plans to support something similar in a different way. I have no issue pushing this configuration (or another) up myself. --- I went to update the `scripts/seed.ts` file that gets generated with a Redwood project and was greeted with a TypeScript error: <details> <summary>TypeScript resolution error</summary> ![image](https://user-images.githubusercontent.com/25166787/142083693-dfd29c6c-b4d1-4353-9cab-731c949bd912.png) </details> The fix is easy enough, here is the `tsconfig.json` file I added to the `scripts/` directory. <details> <summary>scripts/tsconfig.json</summary> ```YAML { "compilerOptions": { "noEmit": true, "allowJs": true, "esModuleInterop": true, "target": "esnext", "module": "esnext", "moduleResolution": "node", "resolveJsonModule": true, "baseUrl": "./", "paths": { "api/*": ["../api/*"], "web/*": ["../web/*"] }, "typeRoots": ["../node_modules/@types", "./node_modules/@types"], "types": ["jest"], }, "include": [ "./", "../.redwood/types/includes/all-*", ] } ``` </details> Was having the `$` character before the resolution something that was desired/necessary? It felt more natural and consistent with each side to leave it off.
process
add a scripts tsconfig json to the create redwood app template hello making this as an issue in the off chance there are plans to support something similar in a different way i have no issue pushing this configuration or another up myself i went to update the scripts seed ts file that gets generated with a redwood project and was greeted with a typescript error typescript resolution error the fix is easy enough here is the tsconfig json file i added to the scripts directory scripts tsconfig json yaml compileroptions noemit true allowjs true esmoduleinterop true target esnext module esnext moduleresolution node resolvejsonmodule true baseurl paths api web typeroots types include redwood types includes all was having the character before the resolution something that was desired necessary it felt more natural and consistent with each side to leave it off
1
11,583
14,444,855,042
IssuesEvent
2020-12-07 21:55:52
googleapis/java-core
https://api.github.com/repos/googleapis/java-core
closed
Please increase google-oauth due to security issue
type: process
We use snyk tool to report security issues in our team. This tools reported us that lib `google-oauth-client` used by `google-cloud-bigquery` has vulnerability. Full snyk info: ``` Overview com.google.oauth-client:google-oauth-client is a powerful and easy-to-use Java library for the OAuth 1.0a and OAuth 2.0 authorization standards. Affected versions of this package are vulnerable to Improper Authorization. PKCE support is not implemented in accordance with the RFC for OAuth 2.0 for Native Apps. Without the use of PKCE, the authorization code returned by an authorization server is not enough to guarantee that the client that issued the initial authorization request is the one that will be authorized. An attacker is able to obtain the authorization code using a malicious app on the client-side and use it to gain authorization to the protected resource. Remediation Upgrade com.google.oauth-client:google-oauth-client to version 1.31.0 or higher. ``` Could you please use newer version of this lib?
1.0
Please increase google-oauth due to security issue - We use snyk tool to report security issues in our team. This tools reported us that lib `google-oauth-client` used by `google-cloud-bigquery` has vulnerability. Full snyk info: ``` Overview com.google.oauth-client:google-oauth-client is a powerful and easy-to-use Java library for the OAuth 1.0a and OAuth 2.0 authorization standards. Affected versions of this package are vulnerable to Improper Authorization. PKCE support is not implemented in accordance with the RFC for OAuth 2.0 for Native Apps. Without the use of PKCE, the authorization code returned by an authorization server is not enough to guarantee that the client that issued the initial authorization request is the one that will be authorized. An attacker is able to obtain the authorization code using a malicious app on the client-side and use it to gain authorization to the protected resource. Remediation Upgrade com.google.oauth-client:google-oauth-client to version 1.31.0 or higher. ``` Could you please use newer version of this lib?
process
please increase google oauth due to security issue we use snyk tool to report security issues in our team this tools reported us that lib google oauth client used by google cloud bigquery has vulnerability full snyk info overview com google oauth client google oauth client is a powerful and easy to use java library for the oauth and oauth authorization standards affected versions of this package are vulnerable to improper authorization pkce support is not implemented in accordance with the rfc for oauth for native apps without the use of pkce the authorization code returned by an authorization server is not enough to guarantee that the client that issued the initial authorization request is the one that will be authorized an attacker is able to obtain the authorization code using a malicious app on the client side and use it to gain authorization to the protected resource remediation upgrade com google oauth client google oauth client to version or higher could you please use newer version of this lib
1
16,017
20,188,226,882
IssuesEvent
2022-02-11 01:19:42
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Configure web apps to reuse authentication tokens securely and handle them like other credentials
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Authentication and authorization
<a href="https://docs.microsoft.com/azure/active-directory/develop/msal-acquire-cache-tokens">Configure web apps to reuse authentication tokens securely and handle them like other credentials</a> <p><b>Why Consider This?</b></p> MSAL caches a token after acquisition, and application code should first try to reuse a token silently from the cache before attempting to acquire one by other means, to improve performance. <p><b>Context</b></p> <p><span>Access tokens enable clients to securely call web APIs protected by Azure. There are several ways to acquire a token by using the Microsoft Authentication Library (MSAL). Some require user interaction through a web browser, while others don't require user interaction. In general, the method used for acquiring a token depends on whether the application is a public client application like desktop or mobile app, or a confidential client application like web app, web API, or daemon application.</span></p><p><span>MSAL caches a token after it's been acquired. Your application code should first try to get a token silently from the cache before attempting to acquire a token by other means.</span></p> <p><b>Suggested Actions</b></p> <p><span>Evaluate web app authentication flow and determine if there's an opportunity to alter behavior to reuse tokens to improve performance.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/active-directory/develop/msal-acquire-cache-tokens" target="_blank"><span>Acquire and cache tokens using MSAL</span></a><span /></p>
1.0
Configure web apps to reuse authentication tokens securely and handle them like other credentials - <a href="https://docs.microsoft.com/azure/active-directory/develop/msal-acquire-cache-tokens">Configure web apps to reuse authentication tokens securely and handle them like other credentials</a> <p><b>Why Consider This?</b></p> MSAL caches a token after acquisition, and application code should first try to reuse a token silently from the cache before attempting to acquire one by other means, to improve performance. <p><b>Context</b></p> <p><span>Access tokens enable clients to securely call web APIs protected by Azure. There are several ways to acquire a token by using the Microsoft Authentication Library (MSAL). Some require user interaction through a web browser, while others don't require user interaction. In general, the method used for acquiring a token depends on whether the application is a public client application like desktop or mobile app, or a confidential client application like web app, web API, or daemon application.</span></p><p><span>MSAL caches a token after it's been acquired. Your application code should first try to get a token silently from the cache before attempting to acquire a token by other means.</span></p> <p><b>Suggested Actions</b></p> <p><span>Evaluate web app authentication flow and determine if there's an opportunity to alter behavior to reuse tokens to improve performance.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/active-directory/develop/msal-acquire-cache-tokens" target="_blank"><span>Acquire and cache tokens using MSAL</span></a><span /></p>
process
configure web apps to reuse authentication tokens securely and handle them like other credentials why consider this msal caches a token after acquisition and application code should first try to reuse a token silently from the cache before attempting to acquire one by other means to improve performance context access tokens enable clients to securely call web apis protected by azure there are several ways to acquire a token by using the microsoft authentication library msal some require user interaction through a web browser while others don t require user interaction in general the method used for acquiring a token depends on whether the application is a public client application like desktop or mobile app or a confidential client application like web app web api or daemon application msal caches a token after it s been acquired your application code should first try to get a token silently from the cache before attempting to acquire a token by other means suggested actions evaluate web app authentication flow and determine if there s an opportunity to alter behavior to reuse tokens to improve performance learn more acquire and cache tokens using msal
1
5,366
8,197,288,150
IssuesEvent
2018-08-31 12:57:31
scieloorg/opac_proc
https://api.github.com/repos/scieloorg/opac_proc
closed
Adicionar o novo campo ``aop_pid`` no processamento
Melhoria Processamento
Os artigos que foram **Ahead** tem no campo **v881** o **PID** antigo do Ahead. Precisamos desse campo no OPAC_PROC para que esteja disponível no Site, com o objetivo de deixar esse **PID** acessíveis. O método no Xylose para obter esse campo é: ``publisher_ahead_id``. Ainda nessa atividade é necessário identificar que foi Ahead e realizar o reprocessamento desses **PIDs** **IMPORTANTE**: Gerar uma lista com esses **PIDS**, para que possamos validar.
1.0
Adicionar o novo campo ``aop_pid`` no processamento - Os artigos que foram **Ahead** tem no campo **v881** o **PID** antigo do Ahead. Precisamos desse campo no OPAC_PROC para que esteja disponível no Site, com o objetivo de deixar esse **PID** acessíveis. O método no Xylose para obter esse campo é: ``publisher_ahead_id``. Ainda nessa atividade é necessário identificar que foi Ahead e realizar o reprocessamento desses **PIDs** **IMPORTANTE**: Gerar uma lista com esses **PIDS**, para que possamos validar.
process
adicionar o novo campo aop pid no processamento os artigos que foram ahead tem no campo o pid antigo do ahead precisamos desse campo no opac proc para que esteja disponível no site com o objetivo de deixar esse pid acessíveis o método no xylose para obter esse campo é publisher ahead id ainda nessa atividade é necessário identificar que foi ahead e realizar o reprocessamento desses pids importante gerar uma lista com esses pids para que possamos validar
1
19,778
26,162,695,159
IssuesEvent
2022-12-31 20:41:27
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
closed
Extract category keywords from dataset title and description
data processing back end
**Is your feature request related to a problem? Please describe.** Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system in merge_data.py. **Describe the solution you'd like** - Combine dataset title and description into single string/ text body - tokenise and remove stopwords - for each remaining keyword, get frequency count in body and return a matching category - retain top 5 most common categories (based on frequency counts) and set as dataset categories **Describe alternatives you've considered** - consider standardising casing and stemming for more accurate comparison - consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough. - If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised. - consider TF-IDF principles (although may be more appropriate at #172 stage) **Additional context** Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
1.0
Extract category keywords from dataset title and description - **Is your feature request related to a problem? Please describe.** Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system in merge_data.py. **Describe the solution you'd like** - Combine dataset title and description into single string/ text body - tokenise and remove stopwords - for each remaining keyword, get frequency count in body and return a matching category - retain top 5 most common categories (based on frequency counts) and set as dataset categories **Describe alternatives you've considered** - consider standardising casing and stemming for more accurate comparison - consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough. - If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised. - consider TF-IDF principles (although may be more appropriate at #172 stage) **Additional context** Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
process
extract category keywords from dataset title and description is your feature request related to a problem please describe too many datasets are set as uncategorised this is because current categorisation uses category keywords provided by the publisher where there are no keywords provided by the publisher the dataset cannot be further categorised by us in the context of the ods catalogue there is a similar ticket but it is a large ticket to tackle this ticket is one step down a subset just to extract keywords out of the dataset title and description to use for categorisation categorisation still using the existing keyword matching system in merge data py describe the solution you d like combine dataset title and description into single string text body tokenise and remove stopwords for each remaining keyword get frequency count in body and return a matching category retain top most common categories based on frequency counts and set as dataset categories describe alternatives you ve considered consider standardising casing and stemming for more accurate comparison consider that there may be no need to cap the number of categories i e returning all categories may be appropriate see this as a of total categories if most categories return most of the time then it s a meaningless solution but because we manually curate category keywords it might actually be filtered enough if the resulting categorisation takes too long compared to current merge data py performance then consider categorising on top n keywords only instead of all keywords the catch is that the top common words may not be useful keywords but we may be satisfied enough with speed and small of datasets left uncategorised consider tf idf principles although may be more appropriate at stage additional context completion of this ticket leaves to be an exploratory piece using unsupervised learning but still a step up in performance until then
1
108,675
9,321,062,170
IssuesEvent
2019-03-27 02:08:51
Softmotions/ejdb
https://api.github.com/repos/Softmotions/ejdb
opened
Invalid state of apply
bug ejdb2 test case required
``` > k add c1 {"n":"a"} < k 1 > k add c1 {"n":"b"} < k 2 > k add c1 {"n":"c"} < k 3 > k del c1 2 < k 2 > k query c1 /* < k 3 {"n":"c"} < k 1 {"n":"a"} < k > k query c1 /* | apply {} < k 3 {"n":"c"} < k ERROR: Invalid JBL buffer (JBL_ERROR_INVALID_BUFFER) ```
1.0
Invalid state of apply - ``` > k add c1 {"n":"a"} < k 1 > k add c1 {"n":"b"} < k 2 > k add c1 {"n":"c"} < k 3 > k del c1 2 < k 2 > k query c1 /* < k 3 {"n":"c"} < k 1 {"n":"a"} < k > k query c1 /* | apply {} < k 3 {"n":"c"} < k ERROR: Invalid JBL buffer (JBL_ERROR_INVALID_BUFFER) ```
non_process
invalid state of apply k add n a k k add n b k k add n c k k del k k query k n c k n a k k query apply k n c k error invalid jbl buffer jbl error invalid buffer
0
15,662
19,847,044,578
IssuesEvent
2022-01-21 07:59:50
ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_non_directional_recovered
https://api.github.com/repos/ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_non_directional_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:59:49.879069. ## Details Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_non_directional_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:59:49.879069. ## Details Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_non_directional_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host wavss a dcl non directional recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
970
3,423,108,273
IssuesEvent
2015-12-09 03:31:46
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
closed
Fix the translation method
bug duplicate Processing
The method that currently tranlates the joystick value does it weirdly so you can often get odd outputs. It should go at the maximum possible thrust (outer circle) if possible.
1.0
Fix the translation method - The method that currently tranlates the joystick value does it weirdly so you can often get odd outputs. It should go at the maximum possible thrust (outer circle) if possible.
process
fix the translation method the method that currently tranlates the joystick value does it weirdly so you can often get odd outputs it should go at the maximum possible thrust outer circle if possible
1
175,457
21,313,819,800
IssuesEvent
2022-04-16 01:02:50
little-apps/SerializableModel
https://api.github.com/repos/little-apps/SerializableModel
opened
CVE-2021-43503 (High) detected in laravel/framework-v5.5.49
security vulnerability
## CVE-2021-43503 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.49</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/a81f23d0ccd2fefa7fa9b79649ab23811631d9bf">https://api.github.com/repos/laravel/framework/zipball/a81f23d0ccd2fefa7fa9b79649ab23811631d9bf</a></p> <p> Dependency Hierarchy: - orchestra/testbench-v3.5.5 (Root Library) - :x: **laravel/framework-v5.5.49** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Remote Code Execution (RCE) vulnerability exists in h laravel 5.8.38 via an unserialize pop chain in (1) __destruct in \Routing\PendingResourceRegistration.php, (2) __cal in Queue\Capsule\Manager.php, and (3) __invoke in mockery\library\Mockery\ClosureWrapper.php. <p>Publish Date: 2022-04-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43503>CVE-2021-43503</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-43503 (High) detected in laravel/framework-v5.5.49 - ## CVE-2021-43503 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.49</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/a81f23d0ccd2fefa7fa9b79649ab23811631d9bf">https://api.github.com/repos/laravel/framework/zipball/a81f23d0ccd2fefa7fa9b79649ab23811631d9bf</a></p> <p> Dependency Hierarchy: - orchestra/testbench-v3.5.5 (Root Library) - :x: **laravel/framework-v5.5.49** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Remote Code Execution (RCE) vulnerability exists in h laravel 5.8.38 via an unserialize pop chain in (1) __destruct in \Routing\PendingResourceRegistration.php, (2) __cal in Queue\Capsule\Manager.php, and (3) __invoke in mockery\library\Mockery\ClosureWrapper.php. <p>Publish Date: 2022-04-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43503>CVE-2021-43503</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in laravel framework cve high severity vulnerability vulnerable library laravel framework the laravel framework library home page a href dependency hierarchy orchestra testbench root library x laravel framework vulnerable library vulnerability details a remote code execution rce vulnerability exists in h laravel via an unserialize pop chain in destruct in routing pendingresourceregistration php cal in queue capsule manager php and invoke in mockery library mockery closurewrapper php publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
202,821
15,858,734,147
IssuesEvent
2021-04-08 07:07:52
MarketSquare/robotframework-robocop
https://api.github.com/repos/MarketSquare/robotframework-robocop
opened
Show available values for configurable rule parameters
documentation enhancement
The only way to get information what values are accepted for a configurable now it to provide incorrect value and get an exception with such information. I would like to have available values listed when calling `-lc / --list-configurables` option and also have a default value marked.
1.0
Show available values for configurable rule parameters - The only way to get information what values are accepted for a configurable now it to provide incorrect value and get an exception with such information. I would like to have available values listed when calling `-lc / --list-configurables` option and also have a default value marked.
non_process
show available values for configurable rule parameters the only way to get information what values are accepted for a configurable now it to provide incorrect value and get an exception with such information i would like to have available values listed when calling lc list configurables option and also have a default value marked
0
14,088
16,978,851,333
IssuesEvent
2021-06-30 05:50:50
open-telemetry/opentelemetry-collector
https://api.github.com/repos/open-telemetry/opentelemetry-collector
opened
Implement redesigned Collector Metrics Processor
area:processor release:required-for-ga spec:metrics
The metrics processor redesign discussion is in progress. See related issue #3185 for proposed design.
1.0
Implement redesigned Collector Metrics Processor - The metrics processor redesign discussion is in progress. See related issue #3185 for proposed design.
process
implement redesigned collector metrics processor the metrics processor redesign discussion is in progress see related issue for proposed design
1
232,605
18,891,606,428
IssuesEvent
2021-11-15 13:49:53
jetstack/cert-manager
https://api.github.com/repos/jetstack/cert-manager
closed
Add upgrade testing
kind/feature priority/important-soon lifecycle/rotten area/testing Epic
**Is your feature request related to a problem? Please describe.** We would want to test the upgrade experience of cert-manager with the v1.0 promise. More specifically to ensure that the upgrade doesn't cause things like existing resources to break. **Describe the solution you'd like** An automated end-to-end style test that we can run as periodic and maybe also on PR that takes the previous version of cert-manager or maybe also making this configurable to test upgrading from a certain release. This version has to be then set up and used for some tests to have cert-manager resourced in the cluster, would also be good to have some in a failing or pending state to ensure that an upgrade doesn't stop any mid-issuance from causing the process to break. An upgrade to the current branch should then be executed and verified. How exactly to be doing this should be in the scope of this work :wink: /kind feature /area testi /milestone v1.1 /priority important-soon
1.0
Add upgrade testing - **Is your feature request related to a problem? Please describe.** We would want to test the upgrade experience of cert-manager with the v1.0 promise. More specifically to ensure that the upgrade doesn't cause things like existing resources to break. **Describe the solution you'd like** An automated end-to-end style test that we can run as periodic and maybe also on PR that takes the previous version of cert-manager or maybe also making this configurable to test upgrading from a certain release. This version has to be then set up and used for some tests to have cert-manager resourced in the cluster, would also be good to have some in a failing or pending state to ensure that an upgrade doesn't stop any mid-issuance from causing the process to break. An upgrade to the current branch should then be executed and verified. How exactly to be doing this should be in the scope of this work :wink: /kind feature /area testi /milestone v1.1 /priority important-soon
non_process
add upgrade testing is your feature request related to a problem please describe we would want to test the upgrade experience of cert manager with the promise more specifically to ensure that the upgrade doesn t cause things like existing resources to break describe the solution you d like an automated end to end style test that we can run as periodic and maybe also on pr that takes the previous version of cert manager or maybe also making this configurable to test upgrading from a certain release this version has to be then set up and used for some tests to have cert manager resourced in the cluster would also be good to have some in a failing or pending state to ensure that an upgrade doesn t stop any mid issuance from causing the process to break an upgrade to the current branch should then be executed and verified how exactly to be doing this should be in the scope of this work wink kind feature area testi milestone priority important soon
0
74,331
7,398,060,436
IssuesEvent
2018-03-19 03:36:23
mono/mono
https://api.github.com/repos/mono/mono
opened
Improve profiler stress test support on pull requests
area-Profiler test enhancement
In particular: * All the same platforms that run the stress tests periodically should be available to trigger on PRs. * `build profiler stress` should trigger builds for all platforms. * `build profiler stress <platform>` should trigger a build for a particular platform.
1.0
Improve profiler stress test support on pull requests - In particular: * All the same platforms that run the stress tests periodically should be available to trigger on PRs. * `build profiler stress` should trigger builds for all platforms. * `build profiler stress <platform>` should trigger a build for a particular platform.
non_process
improve profiler stress test support on pull requests in particular all the same platforms that run the stress tests periodically should be available to trigger on prs build profiler stress should trigger builds for all platforms build profiler stress should trigger a build for a particular platform
0
327,678
24,147,520,214
IssuesEvent
2022-09-21 20:14:10
fga-eps-mds/Cebraspe-Tracker
https://api.github.com/repos/fga-eps-mds/Cebraspe-Tracker
closed
Atualizar Product Backlog
documentation
# Descrição Atualizar o product backlog com alterações definidas # Tarefas - [ ] Adicionar funcionalidades novas - [ ] Remover as obsoletas # Critérios de aceitação - [ ] Deve estar em dia com o que for definido
1.0
Atualizar Product Backlog - # Descrição Atualizar o product backlog com alterações definidas # Tarefas - [ ] Adicionar funcionalidades novas - [ ] Remover as obsoletas # Critérios de aceitação - [ ] Deve estar em dia com o que for definido
non_process
atualizar product backlog descrição atualizar o product backlog com alterações definidas tarefas adicionar funcionalidades novas remover as obsoletas critérios de aceitação deve estar em dia com o que for definido
0
346,332
10,410,971,306
IssuesEvent
2019-09-13 12:49:50
AY1920S1-CS2113T-W17-1/main
https://api.github.com/repos/AY1920S1-CS2113T-W17-1/main
closed
A-Gradle
priority.High
To implement: Gradle with all dependencies & plugin found in https://github.com/AY1920S1-CS2113T-W17-1/main/blob/master/tutorials/gradleTutorial.md#introduction Plugin - CheckStyle - Shadow - JavaFX Dependencies - JUnit
1.0
A-Gradle - To implement: Gradle with all dependencies & plugin found in https://github.com/AY1920S1-CS2113T-W17-1/main/blob/master/tutorials/gradleTutorial.md#introduction Plugin - CheckStyle - Shadow - JavaFX Dependencies - JUnit
non_process
a gradle to implement gradle with all dependencies plugin found in plugin checkstyle shadow javafx dependencies junit
0
2,959
5,955,905,569
IssuesEvent
2017-05-28 11:45:33
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Update documentation for options.insertModal*
inprocess
The example shown [here](http://allenfang.github.io/react-bootstrap-table/example.html#custom) lists in `options` `insertModalBody` `insertModalFooter` However, the docs page for [options](http://allenfang.github.io/react-bootstrap-table/docs.html#options) is missing these. It would be nice to see the documentation for these options also.
1.0
Update documentation for options.insertModal* - The example shown [here](http://allenfang.github.io/react-bootstrap-table/example.html#custom) lists in `options` `insertModalBody` `insertModalFooter` However, the docs page for [options](http://allenfang.github.io/react-bootstrap-table/docs.html#options) is missing these. It would be nice to see the documentation for these options also.
process
update documentation for options insertmodal the example shown lists in options insertmodalbody insertmodalfooter however the docs page for is missing these it would be nice to see the documentation for these options also
1
157,490
12,377,112,978
IssuesEvent
2020-05-19 08:29:28
spring-projects/spring-batch
https://api.github.com/repos/spring-projects/spring-batch
closed
JobLauncherTestUtils doesn't always launch new job instance because generated job parameters aren't unique [BATCH-2783]
in: test type: bug
**[saikatbhadra](https://jira.spring.io/secure/ViewProfile.jspa?name=saikatbhadra)** opened **[BATCH-2783](https://jira.spring.io/browse/BATCH-2783?redirect=false)** and commented We noticed that our integration tests for a batch job (with many tests) were failing intermittently.  After debugging, I found that JobLauncherTestUtils.launchJob was occasionally relaunching the same job instance. After looking at the code, I saw that a random number generator was being used which does not guarantee that the number generate would be unique. I believe a UUID would be more appropriate as strongly guarantees that an ID does not repeat (except in highly exceptional cases).  Pull request here: #3701 --- No further details from [BATCH-2783](https://jira.spring.io/browse/BATCH-2783?redirect=false)
1.0
JobLauncherTestUtils doesn't always launch new job instance because generated job parameters aren't unique [BATCH-2783] - **[saikatbhadra](https://jira.spring.io/secure/ViewProfile.jspa?name=saikatbhadra)** opened **[BATCH-2783](https://jira.spring.io/browse/BATCH-2783?redirect=false)** and commented We noticed that our integration tests for a batch job (with many tests) were failing intermittently.  After debugging, I found that JobLauncherTestUtils.launchJob was occasionally relaunching the same job instance. After looking at the code, I saw that a random number generator was being used which does not guarantee that the number generate would be unique. I believe a UUID would be more appropriate as strongly guarantees that an ID does not repeat (except in highly exceptional cases).  Pull request here: #3701 --- No further details from [BATCH-2783](https://jira.spring.io/browse/BATCH-2783?redirect=false)
non_process
joblaunchertestutils doesn t always launch new job instance because generated job parameters aren t unique opened and commented we noticed that our integration tests for a batch job with many tests were failing intermittently   after debugging i found that joblaunchertestutils launchjob was occasionally relaunching the same job instance after looking at the code i saw that a random number generator was being used which does not guarantee that the number generate would be unique i believe a uuid would be more appropriate as strongly guarantees that an id does not repeat except in highly exceptional cases   pull request here   no further details from
0
6,343
9,380,393,050
IssuesEvent
2019-04-04 16:58:19
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
Re-pin oauth2 for new urls and cut new release
type: process
Token endpoints were recently changed in https://go-review.googlesource.com/c/oauth2/+/136356. Per https://github.com/googleapis/google-cloud-common/issues/260#issuecomment-467669448, let's let this bake for a week or two and then re-pin oauth2 here and in google-api-go-client, and then also re-pin google-api-go-client here. Then, we should release a new version.
1.0
Re-pin oauth2 for new urls and cut new release - Token endpoints were recently changed in https://go-review.googlesource.com/c/oauth2/+/136356. Per https://github.com/googleapis/google-cloud-common/issues/260#issuecomment-467669448, let's let this bake for a week or two and then re-pin oauth2 here and in google-api-go-client, and then also re-pin google-api-go-client here. Then, we should release a new version.
process
re pin for new urls and cut new release token endpoints were recently changed in per let s let this bake for a week or two and then re pin here and in google api go client and then also re pin google api go client here then we should release a new version
1
3,454
6,543,041,760
IssuesEvent
2017-09-02 16:30:49
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
closed
Relase 0.12.1 (RxJava2.x only)
release process RxJava2.x
**Initial release notes**: Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219. **Things to do**: - [x] RxJava2.x branch: - [x] update JavaDoc on `gh-pages` - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Relase 0.12.1 (RxJava2.x only) - **Initial release notes**: Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219. **Things to do**: - [x] RxJava2.x branch: - [x] update JavaDoc on `gh-pages` - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
process
relase x only initial release notes fixed memory leak in prelollipopnetworkobservingstrategy during disposing of an observable issue things to do x branch update javadoc on gh pages bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
1
3,414
6,523,967,818
IssuesEvent
2017-08-29 10:46:03
w3c/w3process
https://api.github.com/repos/w3c/w3process
closed
Requirements for republishing a CR with non-substantive changes
Active Process2018Candidate
Transferred from https://www.w3.org/community/w3process/track/issues/157 State: Open
1.0
Requirements for republishing a CR with non-substantive changes - Transferred from https://www.w3.org/community/w3process/track/issues/157 State: Open
process
requirements for republishing a cr with non substantive changes transferred from state open
1
57,714
7,091,712,176
IssuesEvent
2018-01-12 14:09:21
CityOfPhiladelphia/parks-rec-finder
https://api.github.com/repos/CityOfPhiladelphia/parks-rec-finder
opened
Images for Locations and Things To Do need to be at 25% so that the labels are accessible and readable
design
<img width="1008" alt="screen shot 2018-01-12 at 9 06 57 am" src="https://user-images.githubusercontent.com/16670068/34878462-440401e4-f778-11e7-9147-4eeb49332457.png"> [Spec. Ref. v1.0.4](https://drive.google.com/file/d/15CBYAYFlovs7OOkizQJogSDNMTYamAAJ/view) | URL | App Version | Browser (version) - Size | Device ---------- | --- | ----------- | ------------------------ | ------ pg13.1 | /#/programs/athletic | 0.1.0 | Chrome - 1024 x 768 | Desktop ### Expected Behavior ### Observed Behavior > Steps to Reproduce 1. Opened page 1. Typed into search Bar 1. Cicked Marder #### Proposed Solution ### Screenshot 1.A
1.0
Images for Locations and Things To Do need to be at 25% so that the labels are accessible and readable - <img width="1008" alt="screen shot 2018-01-12 at 9 06 57 am" src="https://user-images.githubusercontent.com/16670068/34878462-440401e4-f778-11e7-9147-4eeb49332457.png"> [Spec. Ref. v1.0.4](https://drive.google.com/file/d/15CBYAYFlovs7OOkizQJogSDNMTYamAAJ/view) | URL | App Version | Browser (version) - Size | Device ---------- | --- | ----------- | ------------------------ | ------ pg13.1 | /#/programs/athletic | 0.1.0 | Chrome - 1024 x 768 | Desktop ### Expected Behavior ### Observed Behavior > Steps to Reproduce 1. Opened page 1. Typed into search Bar 1. Cicked Marder #### Proposed Solution ### Screenshot 1.A
non_process
images for locations and things to do need to be at so that the labels are accessible and readable img width alt screen shot at am src url app version browser version size device programs athletic chrome x desktop expected behavior observed behavior steps to reproduce opened page typed into search bar cicked marder proposed solution screenshot a
0
210,171
16,357,647,460
IssuesEvent
2021-05-14 02:33:59
MarlinFirmware/Marlin
https://api.github.com/repos/MarlinFirmware/Marlin
closed
[BUG] Z Probe Wizard doesn't work without Z_SAFE_HOMING
C: Documentation F: Calibration stale-closing-soon
### Bug Description Z Offset Wizard doesn't work without Z_SAFE_HOMING (e.g. when using Z-endstop for homing, and Z-probe only for bed leveling). ### Configuration Files [configuration.zip](https://github.com/MarlinFirmware/Marlin/files/6087089/configuration.zip) **Required:** Include a ZIP file containing `Configuration.h` and `Configuration_adv.h`. ### Steps to Reproduce 1. Enter menu Configuration 2. Go to Advanced Settings 3. Go to Probe Offsets 4. Go to Z Probe Wizard **Expected behavior:** Printer should either allow to adjust Z Probe Offset at the current position (without Auto Homing), or Auto Home and then move to the middle of the build plate (force behavior like in Z_SAFE_HOMING for Z Probe Wizard). **Actual behavior:** Printer Auto Homes to the bottom left corner of the build plate, and enters Z Offset adjusting mode, with the probe (eg. BL-Touch) outside the build plate, and no option to move the head in any other direction than Z axis.
1.0
[BUG] Z Probe Wizard doesn't work without Z_SAFE_HOMING - ### Bug Description Z Offset Wizard doesn't work without Z_SAFE_HOMING (e.g. when using Z-endstop for homing, and Z-probe only for bed leveling). ### Configuration Files [configuration.zip](https://github.com/MarlinFirmware/Marlin/files/6087089/configuration.zip) **Required:** Include a ZIP file containing `Configuration.h` and `Configuration_adv.h`. ### Steps to Reproduce 1. Enter menu Configuration 2. Go to Advanced Settings 3. Go to Probe Offsets 4. Go to Z Probe Wizard **Expected behavior:** Printer should either allow to adjust Z Probe Offset at the current position (without Auto Homing), or Auto Home and then move to the middle of the build plate (force behavior like in Z_SAFE_HOMING for Z Probe Wizard). **Actual behavior:** Printer Auto Homes to the bottom left corner of the build plate, and enters Z Offset adjusting mode, with the probe (eg. BL-Touch) outside the build plate, and no option to move the head in any other direction than Z axis.
non_process
z probe wizard doesn t work without z safe homing bug description z offset wizard doesn t work without z safe homing e g when using z endstop for homing and z probe only for bed leveling configuration files required include a zip file containing configuration h and configuration adv h steps to reproduce enter menu configuration go to advanced settings go to probe offsets go to z probe wizard expected behavior printer should either allow to adjust z probe offset at the current position without auto homing or auto home and then move to the middle of the build plate force behavior like in z safe homing for z probe wizard actual behavior printer auto homes to the bottom left corner of the build plate and enters z offset adjusting mode with the probe eg bl touch outside the build plate and no option to move the head in any other direction than z axis
0
64,164
14,657,460,642
IssuesEvent
2020-12-28 15:38:47
fu1771695yongxie/yarn
https://api.github.com/repos/fu1771695yongxie/yarn
opened
CVE-2018-1107 (High) detected in https://source.codeaurora.org/quic/le/platform/external/node/v0.10.44
security vulnerability
## CVE-2018-1107 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/quic/le/platform/external/node/v0.10.44</b></p></summary> <p> <p>Library home page: <a href=https://source.codeaurora.org/quic/le/platform/external/node/>https://source.codeaurora.org/quic/le/platform/external/node/</a></p> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> is-my-json-valid before 1.4.2 and 2.17.2 is vulnerable to ReDoS. This library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1546357">https://bugzilla.redhat.com/show_bug.cgi?id=1546357</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 1.4.2,2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1107 (High) detected in https://source.codeaurora.org/quic/le/platform/external/node/v0.10.44 - ## CVE-2018-1107 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/quic/le/platform/external/node/v0.10.44</b></p></summary> <p> <p>Library home page: <a href=https://source.codeaurora.org/quic/le/platform/external/node/>https://source.codeaurora.org/quic/le/platform/external/node/</a></p> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> is-my-json-valid before 1.4.2 and 2.17.2 is vulnerable to ReDoS. This library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1546357">https://bugzilla.redhat.com/show_bug.cgi?id=1546357</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 1.4.2,2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in cve high severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details is my json valid before and is vulnerable to redos this library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
8,569
11,738,855,632
IssuesEvent
2020-03-11 16:43:55
MicrosoftDocs/vsts-docs
https://api.github.com/repos/MicrosoftDocs/vsts-docs
closed
Documentation for "Required Template" approval missing
Pri1 devops-cicd-process/tech devops/prod doc-bug
There is a approval type "Required template", which seems to allow one to require a specific template for an agent. Unfortunately, it's not documented, and it doesn't work as I'd expect it to work. How is this feature supposed to be used? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: b067a175-f640-7503-9c1e-f0130c6dbeda * Version Independent ID: ff743c7b-a103-eae6-4478-62ba995a4b36 * Content: [Pipeline deployment approvals - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/approvals.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @azooinmyluggage * Microsoft Alias: **shashban**
1.0
Documentation for "Required Template" approval missing - There is a approval type "Required template", which seems to allow one to require a specific template for an agent. Unfortunately, it's not documented, and it doesn't work as I'd expect it to work. How is this feature supposed to be used? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: b067a175-f640-7503-9c1e-f0130c6dbeda * Version Independent ID: ff743c7b-a103-eae6-4478-62ba995a4b36 * Content: [Pipeline deployment approvals - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/approvals?view=azure-devops&tabs=check-pass) * Content Source: [docs/pipelines/process/approvals.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/approvals.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @azooinmyluggage * Microsoft Alias: **shashban**
process
documentation for required template approval missing there is a approval type required template which seems to allow one to require a specific template for an agent unfortunately it s not documented and it doesn t work as i d expect it to work how is this feature supposed to be used document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login azooinmyluggage microsoft alias shashban
1
17,202
6,387,273,581
IssuesEvent
2017-08-03 13:20:02
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
opened
Provide nightly wheels?
Build / CI Need Contributor
Using scikit-optimize I realized that it would be great if we could provide nightly wheels. The numpy and scipy wheels mean that we can run CI against their master, which is great for us and them. It would be good if we could provide the same courtesy to our downsteam users. (also I just asked cython to do the same for us, so we should set a good example ;) I have not looked into how scipy and numpy do it, but I'd hope we can steal some of their tools.
1.0
Provide nightly wheels? - Using scikit-optimize I realized that it would be great if we could provide nightly wheels. The numpy and scipy wheels mean that we can run CI against their master, which is great for us and them. It would be good if we could provide the same courtesy to our downsteam users. (also I just asked cython to do the same for us, so we should set a good example ;) I have not looked into how scipy and numpy do it, but I'd hope we can steal some of their tools.
non_process
provide nightly wheels using scikit optimize i realized that it would be great if we could provide nightly wheels the numpy and scipy wheels mean that we can run ci against their master which is great for us and them it would be good if we could provide the same courtesy to our downsteam users also i just asked cython to do the same for us so we should set a good example i have not looked into how scipy and numpy do it but i d hope we can steal some of their tools
0
3,989
6,917,720,900
IssuesEvent
2017-11-29 09:36:04
nerdalize/nerd
https://api.github.com/repos/nerdalize/nerd
opened
Allow testing of the CLI against configurable Kubernetes versions
Dev Process
We want to make the minikube Kubernetes version we test against configurable through a environment variable ## Expected Behavior `KUBE_VERSION=1.8.0; ./make.sh test` should switch to testing against a 1.8.0 minikube vm ## Actual Behavior It tests against a hardcoded Kubernetes version (1.7.5) ## Steps to Reproduce the Problem 1. `./make.sh test` ## Specifications - Version 0.6.0 dev branch - Platform: MacOS - Subsystem: High Sierra ## Anything else we need to know? We are probably not able to provide full support but we should be able to test and document what kube versions are supported
1.0
Allow testing of the CLI against configurable Kubernetes versions - We want to make the minikube Kubernetes version we test against configurable through a environment variable ## Expected Behavior `KUBE_VERSION=1.8.0; ./make.sh test` should switch to testing against a 1.8.0 minikube vm ## Actual Behavior It tests against a hardcoded Kubernetes version (1.7.5) ## Steps to Reproduce the Problem 1. `./make.sh test` ## Specifications - Version 0.6.0 dev branch - Platform: MacOS - Subsystem: High Sierra ## Anything else we need to know? We are probably not able to provide full support but we should be able to test and document what kube versions are supported
process
allow testing of the cli against configurable kubernetes versions we want to make the minikube kubernetes version we test against configurable through a environment variable expected behavior kube version make sh test should switch to testing against a minikube vm actual behavior it tests against a hardcoded kubernetes version steps to reproduce the problem make sh test specifications version dev branch platform macos subsystem high sierra anything else we need to know we are probably not able to provide full support but we should be able to test and document what kube versions are supported
1
21,051
27,994,272,452
IssuesEvent
2023-03-27 07:18:48
Altinn/app-frontend-react
https://api.github.com/repos/Altinn/app-frontend-react
closed
App crashes on confirmation task when user's role is not allowed to confirm
kind/bug org/brg area/process feature-complete area/authorization
### Description of the bug Our case is such that there are 20 roles allowed to fill out the form, but only 13 roles are allowed to perform the confirmation step. However, when setup for this is done in the apps policy file, as described by the documentation, and we test with a user that has one of these 7 roles that can perform the first step but not the second, the app crashes. That is, the user is allowed to fill out the form and hit next, can see the confirmaton page (as expected) but they can also hit 'confirm'. And when they do, the app crashes with a 403 error. This isn't exactly the ideal way to prevent the user from confirming. We'd much rather see some form of error message that explains why the user is not allowed to perform the confirmation. The case here is that you have `read` access, but not `confirm`. ### Steps To Reproduce 1. Have an app with two tasks defined. 2. Set Role1 and Role2 to be able to perform one process task, but only Role1 being able to perform the next (in the policy file). 3. Try to perform Task1 with Role2 and see that it works. Move on to Task2 4. See that you are in Task2 and try to move on passed this with Role2. 5. The app crashes ### Additional Information A proposed quick-fix is for front-end to handle the resulting 403 in a way that makes sense for the user.
1.0
App crashes on confirmation task when user's role is not allowed to confirm - ### Description of the bug Our case is such that there are 20 roles allowed to fill out the form, but only 13 roles are allowed to perform the confirmation step. However, when setup for this is done in the apps policy file, as described by the documentation, and we test with a user that has one of these 7 roles that can perform the first step but not the second, the app crashes. That is, the user is allowed to fill out the form and hit next, can see the confirmaton page (as expected) but they can also hit 'confirm'. And when they do, the app crashes with a 403 error. This isn't exactly the ideal way to prevent the user from confirming. We'd much rather see some form of error message that explains why the user is not allowed to perform the confirmation. The case here is that you have `read` access, but not `confirm`. ### Steps To Reproduce 1. Have an app with two tasks defined. 2. Set Role1 and Role2 to be able to perform one process task, but only Role1 being able to perform the next (in the policy file). 3. Try to perform Task1 with Role2 and see that it works. Move on to Task2 4. See that you are in Task2 and try to move on passed this with Role2. 5. The app crashes ### Additional Information A proposed quick-fix is for front-end to handle the resulting 403 in a way that makes sense for the user.
process
app crashes on confirmation task when user s role is not allowed to confirm description of the bug our case is such that there are roles allowed to fill out the form but only roles are allowed to perform the confirmation step however when setup for this is done in the apps policy file as described by the documentation and we test with a user that has one of these roles that can perform the first step but not the second the app crashes that is the user is allowed to fill out the form and hit next can see the confirmaton page as expected but they can also hit confirm and when they do the app crashes with a error this isn t exactly the ideal way to prevent the user from confirming we d much rather see some form of error message that explains why the user is not allowed to perform the confirmation the case here is that you have read access but not confirm steps to reproduce have an app with two tasks defined set and to be able to perform one process task but only being able to perform the next in the policy file try to perform with and see that it works move on to see that you are in and try to move on passed this with the app crashes additional information a proposed quick fix is for front end to handle the resulting in a way that makes sense for the user
1
5,378
8,204,874,665
IssuesEvent
2018-09-03 08:19:05
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
Alba proxy & maintenance process should automatically pick up changed redis-lru config
process_wontfix type_enhancement
The current solution for lru eviction from the alba fragment cache allows specifying a redis instance to be used. Only 1 ip can be specified. If the host goes down, they may want to configure another redis instance on another host to be used for the lru tracking. This is possible, but afterwards they'll have to restart all related proxy & maintenance processes. This ticket is about eliminating the need for this restart.
1.0
Alba proxy & maintenance process should automatically pick up changed redis-lru config - The current solution for lru eviction from the alba fragment cache allows specifying a redis instance to be used. Only 1 ip can be specified. If the host goes down, they may want to configure another redis instance on another host to be used for the lru tracking. This is possible, but afterwards they'll have to restart all related proxy & maintenance processes. This ticket is about eliminating the need for this restart.
process
alba proxy maintenance process should automatically pick up changed redis lru config the current solution for lru eviction from the alba fragment cache allows specifying a redis instance to be used only ip can be specified if the host goes down they may want to configure another redis instance on another host to be used for the lru tracking this is possible but afterwards they ll have to restart all related proxy maintenance processes this ticket is about eliminating the need for this restart
1
21,490
29,579,318,506
IssuesEvent
2023-06-07 03:32:22
phuocduong-agilityio/internship-huy-dao
https://api.github.com/repos/phuocduong-agilityio/internship-huy-dao
closed
Create `Button` Component
In-process
- [x] Create the `Button` component - [x] Split object into a storybook - [x] Set style component - **NOTE**: - The three component objects I create in the storybook will correspond to: ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/be578005-a3be-4089-adc8-70c985105b62), ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/736c48f3-d798-468f-8ab0-1ea2d0722ddb), ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/6770d5cf-6196-4fb0-bb25-608fa6ac2331) - See more details: [Link](https://docs.google.com/document/d/1iTrT9cdKrMCjUc6aDLS514gEk7BUFj-m2SCs2wCy2Lw/edit#heading=h.7t5pbcmau6ih)
1.0
Create `Button` Component - - [x] Create the `Button` component - [x] Split object into a storybook - [x] Set style component - **NOTE**: - The three component objects I create in the storybook will correspond to: ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/be578005-a3be-4089-adc8-70c985105b62), ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/736c48f3-d798-468f-8ab0-1ea2d0722ddb), ![image](https://github.com/phuocduong-agilityio/internship-huy-dao/assets/85359200/6770d5cf-6196-4fb0-bb25-608fa6ac2331) - See more details: [Link](https://docs.google.com/document/d/1iTrT9cdKrMCjUc6aDLS514gEk7BUFj-m2SCs2wCy2Lw/edit#heading=h.7t5pbcmau6ih)
process
create button component create the button component split object into a storybook set style component note the three component objects i create in the storybook will correspond to see more details
1
316,283
9,640,066,229
IssuesEvent
2019-05-16 14:46:45
INN/largo
https://api.github.com/repos/INN/largo
closed
In search results, apply `overflow-wrap: break-word;` to URLs
priority: normal status: needs review
<img width="389" alt="screen shot 2018-08-07 at 5 19 19 pm" src="https://user-images.githubusercontent.com/1754187/43803455-e2287576-9a66-11e8-9598-891e59d56620.png"> Given a URL whose length between the first character and the first hyphen is longer than the text column is wide, we need to make sure that wraps at the end of the column rather than overflowing. https://github.com/INN/largo/blob/62127383aae0e070883c4f915016582d14191535/partials/content-search.php#L20-L24 - [ ] adding `overflow-wrap: break-word;` to the `a` tag will do it, but it needs to be scoped to the search result markup in [partials/content-search.php](https://github.com/INN/largo/blob/0.5-dev/partials/content-search.php). - [ ] If the URL is always going to be longer than the date, could we also get rid of the `.sep` separator and just have the date and URL be on separate lines?
1.0
In search results, apply `overflow-wrap: break-word;` to URLs - <img width="389" alt="screen shot 2018-08-07 at 5 19 19 pm" src="https://user-images.githubusercontent.com/1754187/43803455-e2287576-9a66-11e8-9598-891e59d56620.png"> Given a URL whose length between the first character and the first hyphen is longer than the text column is wide, we need to make sure that wraps at the end of the column rather than overflowing. https://github.com/INN/largo/blob/62127383aae0e070883c4f915016582d14191535/partials/content-search.php#L20-L24 - [ ] adding `overflow-wrap: break-word;` to the `a` tag will do it, but it needs to be scoped to the search result markup in [partials/content-search.php](https://github.com/INN/largo/blob/0.5-dev/partials/content-search.php). - [ ] If the URL is always going to be longer than the date, could we also get rid of the `.sep` separator and just have the date and URL be on separate lines?
non_process
in search results apply overflow wrap break word to urls img width alt screen shot at pm src given a url whose length between the first character and the first hyphen is longer than the text column is wide we need to make sure that wraps at the end of the column rather than overflowing adding overflow wrap break word to the a tag will do it but it needs to be scoped to the search result markup in if the url is always going to be longer than the date could we also get rid of the sep separator and just have the date and url be on separate lines
0
66,433
20,196,564,945
IssuesEvent
2022-02-11 11:12:25
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
Message bubbles: horizontal black lines between messages (sometimes)
T-Defect
### Steps to reproduce 1. Check any room to which different people send messages to. ### Outcome #### What did you expect? #### What happened instead? Sometimes you can see black horizontal lines spanning across the whole width in between messages of different people but not between two messages of the same person. See attached screenshot: ![20F19FB4-2185-48EE-97D8-E26FAFEE451A](https://user-images.githubusercontent.com/39308834/153581725-4cfff642-b090-4a13-bf2b-46c06ac58256.jpeg) ### Your phone model iPhone 8 ### Operating system version iOS 15.3 ### Application version Element 1.8.0 ### Homeserver Synapse 1.52.0 ### Will you send logs? No
1.0
Message bubbles: horizontal black lines between messages (sometimes) - ### Steps to reproduce 1. Check any room to which different people send messages to. ### Outcome #### What did you expect? #### What happened instead? Sometimes you can see black horizontal lines spanning across the whole width in between messages of different people but not between two messages of the same person. See attached screenshot: ![20F19FB4-2185-48EE-97D8-E26FAFEE451A](https://user-images.githubusercontent.com/39308834/153581725-4cfff642-b090-4a13-bf2b-46c06ac58256.jpeg) ### Your phone model iPhone 8 ### Operating system version iOS 15.3 ### Application version Element 1.8.0 ### Homeserver Synapse 1.52.0 ### Will you send logs? No
non_process
message bubbles horizontal black lines between messages sometimes steps to reproduce check any room to which different people send messages to outcome what did you expect what happened instead sometimes you can see black horizontal lines spanning across the whole width in between messages of different people but not between two messages of the same person see attached screenshot your phone model iphone operating system version ios application version element homeserver synapse will you send logs no
0
62,917
17,262,031,757
IssuesEvent
2021-07-22 08:59:00
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
CDS / CDNSKEY records missing in zone transfers when direct-dnskey=yes
auth defect
- Program: Authoritative - Issue type: Bug report ### Short description When `direct-dnskey=yes` is set, CDS and CDNSKEY records are not included in AXFR zone transfers. ### Environment <!-- Tell us about the environment --> - Operating system: ubuntu 16.04 - Software version: ``` root@ns:~# pdns_server --version Sep 27 05:38:22 PowerDNS Authoritative Server 0.0.2458g628b164 (C) 2001-2018 PowerDNS.COM BV Sep 27 05:38:22 Using 64-bits mode. Built using gcc 5.4.0 20160609 on Sep 25 2018 13:40:42 by root@16deb14fb81c. Sep 27 05:38:22 PowerDNS comes with ABSOLUTELY NO WARRANTY. This is free software, and you are welcome to redistribute it according to the terms of the GPL version 2. Sep 27 05:38:22 Features: sodium openssl lua PKCS#11 Sep 27 05:38:22 Built-in modules: Sep 27 05:38:22 Configured with: " '--host=' '--build=' '--prefix=/usr' '--sysconfdir=/etc/powerdns' '--mandir=${prefix}/share/man' '--infodir=${prefix}/share/info' '--libdir=${prefix}/lib/x86_64-linux-gnu' '--libexecdir=${prefix}/lib' '--with-dynmodules=bind ldap pipe gmysql godbc gpgsql gsqlite3 geoip lua lua2 mydns remote random opendbx tinydns' '--with-modules=' '--enable-tools' '--enable-ixfrdist' '--enable-unit-tests' '--with-lua=luajit' '--enable-experimental-pkcs11' '--enable-systemd' '--with-systemd=/lib/systemd/system' '--enable-libsodium' 'build_alias=' 'host_alias=' 'CFLAGS=-g -O2 -fstack-protector-strong -Wformat -Werror=format-security' 'LDFLAGS=-Wl,-Bsymbolic-functions -Wl,-z,relro' 'CPPFLAGS=-Wdate-time -D_FORTIFY_SOURCE=2' 'CXXFLAGS=-g -O2 -fstack-protector-strong -Wformat -Werror=format-security'" ``` - Software source: PowerDNS repository ### Steps to reproduce 1. In the powerdns configuration file, set `direct-dnskey=yes` and restart the server 2. ```pdnsutil create-zone dnscamel.ch ns.ed25519.ch``` 3. ```pdnsutil set-kind dnscamel.ch master``` 4. ```pdnsutil secure-zone dnscamel.ch``` 5. ```pdnsutil set-publish-cds dnscamel.ch 2``` 6. ```pdnsutil set-publish-cdnskey dnscamel.ch``` 7. ```dig axfr dnscamel.ch @localhost``` ### Expected behaviour The synthesized CDS/CDNSKEY records are included in the AXFR ( which is the case for `direct-dnskey=no` ) ``` root@ns:~# dig axfr dnscamel.ch @localhost ; <<>> DiG 9.10.3-P4-Ubuntu <<>> axfr dnscamel.ch @localhost ;; global options: +cmd dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 dnscamel.ch. 3600 IN RRSIG SOA 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. NLgbB+VY/SoDzWEGExFelSMvcnLaHcNxvFlBh2hl5C0wBjiSTP1PFw5x ObT9NSolTkSCv/qLyY1kDuDFWd0qRA== dnscamel.ch. 3600 IN NS ns.ed25519.ch. dnscamel.ch. 3600 IN RRSIG NS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. zrsuIO4/O/wUYnNS3huTUE1uxjqbytmuoK/zWrf4eoxnAVJlw9bcVcLq vDGg/l0iTfvilE3mmyK1yO7BDxP0yA== dnscamel.ch. 3600 IN CDS 41000 13 2 8FEB053BF16AC31FF1672857206E5A3188A1737B27CA24CC7C547B92 87545F80 dnscamel.ch. 3600 IN RRSIG CDS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. p/ECK9QjUpXJVJtYsO1ncz+1AJfp0l6btCvWaec6AK79G6VL3phB17l1 CkD20Fxz8Q/hJehzQrigAbCtAlLOTg== dnscamel.ch. 3600 IN NSEC dnscamel.ch. NS SOA RRSIG NSEC DNSKEY CDS CDNSKEY dnscamel.ch. 3600 IN RRSIG NSEC 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. vccM6/DxRlnymGrtpVhUvINcHr9ww1wFQrbC7uB+kyYhtRbgAOkPo3o2 rczB1tobW74v0CDfHJ7f0/lMjxHV1g== dnscamel.ch. 3600 IN DNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG DNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. dcMQSS+JsS/4vkauwERvjk7qkfljpTAGAPrlCP+Px0WnKWynvN7chRhM Rrpkk53Lot1x1WdqOfitTHHVRfDyyA== dnscamel.ch. 3600 IN CDNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG CDNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. yKsTe65wBoB0lbfP42X58Y5LFKw8VRuQ6z/ZCVLcX/txS1RMuNcsWieo OsSBJdbytEMMcenuLX01tIofK+sFug== dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 ;; Query time: 9 msec ;; SERVER: 127.0.0.1#53(127.0.0.1) ;; WHEN: Thu Sep 27 05:35:59 CEST 2018 ;; XFR size: 13 records (messages 3, bytes 1152) ``` ### Actual behaviour CDS and CDNSKEY records are not present in the AXFR. ``` root@ns:~# dig axfr dnscamel.ch @localhost ; <<>> DiG 9.10.3-P4-Ubuntu <<>> axfr dnscamel.ch @localhost ;; global options: +cmd dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 dnscamel.ch. 3600 IN RRSIG SOA 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. FB+suE0dbPuq4Sfn8LcFHKBm34HuxBo2eR5kzjSc/IbO4kRFYwGxvE7v Sog4Z0JRZYBMHgs/F79F1zRBb1JtEw== dnscamel.ch. 3600 IN DNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG DNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. JUVB0lvFgp7jyP0N2UICkD5plJH1ZMz399EwYM9u9x83xvNrv5Naetxv c2qGhPnPLqa/2vz98NYX2l3H1kjlBw== dnscamel.ch. 3600 IN NSEC dnscamel.ch. NS SOA RRSIG NSEC DNSKEY dnscamel.ch. 3600 IN RRSIG NSEC 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. TxrHYdTvv91h5lvGRszZgExc55iAFPKoubv01fYixyb/01VOzrYjWpvO AOpsU/IaYrhBoy8l6upyEAHq6Q9OJg== dnscamel.ch. 3600 IN NS ns.ed25519.ch. dnscamel.ch. 3600 IN RRSIG NS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. 0vsU+2B5nLaP/qfYosLafXJ6lryCD2EENpU0azm5fpbA//+xhY1FbAqr +jOlR0Fwkp15bkl2U6muPKqj0LNopw== dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 ;; Query time: 8 msec ;; SERVER: 127.0.0.1#53(127.0.0.1) ;; WHEN: Thu Sep 27 05:34:18 CEST 2018 ;; XFR size: 9 records (messages 3, bytes 809) ```
1.0
CDS / CDNSKEY records missing in zone transfers when direct-dnskey=yes - - Program: Authoritative - Issue type: Bug report ### Short description When `direct-dnskey=yes` is set, CDS and CDNSKEY records are not included in AXFR zone transfers. ### Environment <!-- Tell us about the environment --> - Operating system: ubuntu 16.04 - Software version: ``` root@ns:~# pdns_server --version Sep 27 05:38:22 PowerDNS Authoritative Server 0.0.2458g628b164 (C) 2001-2018 PowerDNS.COM BV Sep 27 05:38:22 Using 64-bits mode. Built using gcc 5.4.0 20160609 on Sep 25 2018 13:40:42 by root@16deb14fb81c. Sep 27 05:38:22 PowerDNS comes with ABSOLUTELY NO WARRANTY. This is free software, and you are welcome to redistribute it according to the terms of the GPL version 2. Sep 27 05:38:22 Features: sodium openssl lua PKCS#11 Sep 27 05:38:22 Built-in modules: Sep 27 05:38:22 Configured with: " '--host=' '--build=' '--prefix=/usr' '--sysconfdir=/etc/powerdns' '--mandir=${prefix}/share/man' '--infodir=${prefix}/share/info' '--libdir=${prefix}/lib/x86_64-linux-gnu' '--libexecdir=${prefix}/lib' '--with-dynmodules=bind ldap pipe gmysql godbc gpgsql gsqlite3 geoip lua lua2 mydns remote random opendbx tinydns' '--with-modules=' '--enable-tools' '--enable-ixfrdist' '--enable-unit-tests' '--with-lua=luajit' '--enable-experimental-pkcs11' '--enable-systemd' '--with-systemd=/lib/systemd/system' '--enable-libsodium' 'build_alias=' 'host_alias=' 'CFLAGS=-g -O2 -fstack-protector-strong -Wformat -Werror=format-security' 'LDFLAGS=-Wl,-Bsymbolic-functions -Wl,-z,relro' 'CPPFLAGS=-Wdate-time -D_FORTIFY_SOURCE=2' 'CXXFLAGS=-g -O2 -fstack-protector-strong -Wformat -Werror=format-security'" ``` - Software source: PowerDNS repository ### Steps to reproduce 1. In the powerdns configuration file, set `direct-dnskey=yes` and restart the server 2. ```pdnsutil create-zone dnscamel.ch ns.ed25519.ch``` 3. ```pdnsutil set-kind dnscamel.ch master``` 4. ```pdnsutil secure-zone dnscamel.ch``` 5. ```pdnsutil set-publish-cds dnscamel.ch 2``` 6. ```pdnsutil set-publish-cdnskey dnscamel.ch``` 7. ```dig axfr dnscamel.ch @localhost``` ### Expected behaviour The synthesized CDS/CDNSKEY records are included in the AXFR ( which is the case for `direct-dnskey=no` ) ``` root@ns:~# dig axfr dnscamel.ch @localhost ; <<>> DiG 9.10.3-P4-Ubuntu <<>> axfr dnscamel.ch @localhost ;; global options: +cmd dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 dnscamel.ch. 3600 IN RRSIG SOA 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. NLgbB+VY/SoDzWEGExFelSMvcnLaHcNxvFlBh2hl5C0wBjiSTP1PFw5x ObT9NSolTkSCv/qLyY1kDuDFWd0qRA== dnscamel.ch. 3600 IN NS ns.ed25519.ch. dnscamel.ch. 3600 IN RRSIG NS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. zrsuIO4/O/wUYnNS3huTUE1uxjqbytmuoK/zWrf4eoxnAVJlw9bcVcLq vDGg/l0iTfvilE3mmyK1yO7BDxP0yA== dnscamel.ch. 3600 IN CDS 41000 13 2 8FEB053BF16AC31FF1672857206E5A3188A1737B27CA24CC7C547B92 87545F80 dnscamel.ch. 3600 IN RRSIG CDS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. p/ECK9QjUpXJVJtYsO1ncz+1AJfp0l6btCvWaec6AK79G6VL3phB17l1 CkD20Fxz8Q/hJehzQrigAbCtAlLOTg== dnscamel.ch. 3600 IN NSEC dnscamel.ch. NS SOA RRSIG NSEC DNSKEY CDS CDNSKEY dnscamel.ch. 3600 IN RRSIG NSEC 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. vccM6/DxRlnymGrtpVhUvINcHr9ww1wFQrbC7uB+kyYhtRbgAOkPo3o2 rczB1tobW74v0CDfHJ7f0/lMjxHV1g== dnscamel.ch. 3600 IN DNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG DNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. dcMQSS+JsS/4vkauwERvjk7qkfljpTAGAPrlCP+Px0WnKWynvN7chRhM Rrpkk53Lot1x1WdqOfitTHHVRfDyyA== dnscamel.ch. 3600 IN CDNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG CDNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. yKsTe65wBoB0lbfP42X58Y5LFKw8VRuQ6z/ZCVLcX/txS1RMuNcsWieo OsSBJdbytEMMcenuLX01tIofK+sFug== dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 ;; Query time: 9 msec ;; SERVER: 127.0.0.1#53(127.0.0.1) ;; WHEN: Thu Sep 27 05:35:59 CEST 2018 ;; XFR size: 13 records (messages 3, bytes 1152) ``` ### Actual behaviour CDS and CDNSKEY records are not present in the AXFR. ``` root@ns:~# dig axfr dnscamel.ch @localhost ; <<>> DiG 9.10.3-P4-Ubuntu <<>> axfr dnscamel.ch @localhost ;; global options: +cmd dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 dnscamel.ch. 3600 IN RRSIG SOA 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. FB+suE0dbPuq4Sfn8LcFHKBm34HuxBo2eR5kzjSc/IbO4kRFYwGxvE7v Sog4Z0JRZYBMHgs/F79F1zRBb1JtEw== dnscamel.ch. 3600 IN DNSKEY 257 3 13 SKazOyrzl7IiqvI6M+UpakwUUWpxk8un2mdnw/mD4boh0K5Fc3S4PQX9 wK6xlGZR7XO+lcRZyl5rhe81bRqWSA== dnscamel.ch. 3600 IN RRSIG DNSKEY 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. JUVB0lvFgp7jyP0N2UICkD5plJH1ZMz399EwYM9u9x83xvNrv5Naetxv c2qGhPnPLqa/2vz98NYX2l3H1kjlBw== dnscamel.ch. 3600 IN NSEC dnscamel.ch. NS SOA RRSIG NSEC DNSKEY dnscamel.ch. 3600 IN RRSIG NSEC 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. TxrHYdTvv91h5lvGRszZgExc55iAFPKoubv01fYixyb/01VOzrYjWpvO AOpsU/IaYrhBoy8l6upyEAHq6Q9OJg== dnscamel.ch. 3600 IN NS ns.ed25519.ch. dnscamel.ch. 3600 IN RRSIG NS 13 2 3600 20181011000000 20180920000000 41000 dnscamel.ch. 0vsU+2B5nLaP/qfYosLafXJ6lryCD2EENpU0azm5fpbA//+xhY1FbAqr +jOlR0Fwkp15bkl2U6muPKqj0LNopw== dnscamel.ch. 3600 IN SOA ns.ed25519.ch. anooky.anooky.ch. 2544 10800 3600 604800 3600 ;; Query time: 8 msec ;; SERVER: 127.0.0.1#53(127.0.0.1) ;; WHEN: Thu Sep 27 05:34:18 CEST 2018 ;; XFR size: 9 records (messages 3, bytes 809) ```
non_process
cds cdnskey records missing in zone transfers when direct dnskey yes program authoritative issue type bug report short description when direct dnskey yes is set cds and cdnskey records are not included in axfr zone transfers environment operating system ubuntu software version root ns pdns server version sep powerdns authoritative server c powerdns com bv sep using bits mode built using gcc on sep by root sep powerdns comes with absolutely no warranty this is free software and you are welcome to redistribute it according to the terms of the gpl version sep features sodium openssl lua pkcs sep built in modules sep configured with host build prefix usr sysconfdir etc powerdns mandir prefix share man infodir prefix share info libdir prefix lib linux gnu libexecdir prefix lib with dynmodules bind ldap pipe gmysql godbc gpgsql geoip lua mydns remote random opendbx tinydns with modules enable tools enable ixfrdist enable unit tests with lua luajit enable experimental enable systemd with systemd lib systemd system enable libsodium build alias host alias cflags g fstack protector strong wformat werror format security ldflags wl bsymbolic functions wl z relro cppflags wdate time d fortify source cxxflags g fstack protector strong wformat werror format security software source powerdns repository steps to reproduce in the powerdns configuration file set direct dnskey yes and restart the server pdnsutil create zone dnscamel ch ns ch pdnsutil set kind dnscamel ch master pdnsutil secure zone dnscamel ch pdnsutil set publish cds dnscamel ch pdnsutil set publish cdnskey dnscamel ch dig axfr dnscamel ch localhost expected behaviour the synthesized cds cdnskey records are included in the axfr which is the case for direct dnskey no root ns dig axfr dnscamel ch localhost dig ubuntu axfr dnscamel ch localhost global options cmd dnscamel ch in soa ns ch anooky anooky ch dnscamel ch in rrsig soa dnscamel ch nlgbb vy dnscamel ch in ns ns ch dnscamel ch in rrsig ns dnscamel ch o vdgg dnscamel ch in cds dnscamel ch in rrsig cds dnscamel ch p hjehzqrigabctallotg dnscamel ch in nsec dnscamel ch ns soa rrsig nsec dnskey cds cdnskey dnscamel ch in rrsig nsec dnscamel ch dnscamel ch in dnskey dnscamel ch in rrsig dnskey dnscamel ch dcmqss jss dnscamel ch in cdnskey dnscamel ch in rrsig cdnskey dnscamel ch zcvlcx sfug dnscamel ch in soa ns ch anooky anooky ch query time msec server when thu sep cest xfr size records messages bytes actual behaviour cds and cdnskey records are not present in the axfr root ns dig axfr dnscamel ch localhost dig ubuntu axfr dnscamel ch localhost global options cmd dnscamel ch in soa ns ch anooky anooky ch dnscamel ch in rrsig soa dnscamel ch fb dnscamel ch in dnskey dnscamel ch in rrsig dnskey dnscamel ch dnscamel ch in nsec dnscamel ch ns soa rrsig nsec dnskey dnscamel ch in rrsig nsec dnscamel ch aopsu dnscamel ch in ns ns ch dnscamel ch in rrsig ns dnscamel ch dnscamel ch in soa ns ch anooky anooky ch query time msec server when thu sep cest xfr size records messages bytes
0
19,409
25,555,428,105
IssuesEvent
2022-11-30 06:01:27
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] [Auth server] Sign in button is disabled in the following screen
Bug P1 Process: Fixed Auth server
Steps: 1. Click on PM URL 2. Without entering the data click on Submit button 3. Enter the required values and observe AR: Sign in button is disabled ER: Sign in button should be enabled in the following scenario ![image](https://user-images.githubusercontent.com/71445210/204558291-54dae9c1-4e37-4646-8b4f-459d62e661fb.png)
1.0
[PM] [Auth server] Sign in button is disabled in the following screen - Steps: 1. Click on PM URL 2. Without entering the data click on Submit button 3. Enter the required values and observe AR: Sign in button is disabled ER: Sign in button should be enabled in the following scenario ![image](https://user-images.githubusercontent.com/71445210/204558291-54dae9c1-4e37-4646-8b4f-459d62e661fb.png)
process
sign in button is disabled in the following screen steps click on pm url without entering the data click on submit button enter the required values and observe ar sign in button is disabled er sign in button should be enabled in the following scenario
1
2,623
5,396,836,773
IssuesEvent
2017-02-27 13:01:58
mesosphere/marathon
https://api.github.com/repos/mesosphere/marathon
closed
#5063 Upload Snapshot Docker Image to Docker Hub
Epic:Improve CI and Release Process in progress ready for review task
Commits to master should be release as `latest-dev` on Docker Hub.
1.0
#5063 Upload Snapshot Docker Image to Docker Hub - Commits to master should be release as `latest-dev` on Docker Hub.
process
upload snapshot docker image to docker hub commits to master should be release as latest dev on docker hub
1
121,264
10,163,551,184
IssuesEvent
2019-08-07 09:32:12
HadiModarres/s3_uploader_service
https://api.github.com/repos/HadiModarres/s3_uploader_service
closed
Logs
bug testing
Please keep a log of succesful uploads and also a log of errors. Create a folder with two files: sucessful_uploads and error successful_uploads keeps a log of successful uploads and from which path a file was taken and at which place in what bucket and region was placed . error will keep a record of any erros that have occured.
1.0
Logs - Please keep a log of succesful uploads and also a log of errors. Create a folder with two files: sucessful_uploads and error successful_uploads keeps a log of successful uploads and from which path a file was taken and at which place in what bucket and region was placed . error will keep a record of any erros that have occured.
non_process
logs please keep a log of succesful uploads and also a log of errors create a folder with two files sucessful uploads and error successful uploads keeps a log of successful uploads and from which path a file was taken and at which place in what bucket and region was placed error will keep a record of any erros that have occured
0
22,138
30,683,446,461
IssuesEvent
2023-07-26 10:40:46
ppy/osu-web
https://api.github.com/repos/ppy/osu-web
closed
Beatmap covers API endpoints return images of the old format for most non-recent graveyard maps
area:beatmap-processing
The size of images returned by the beatmaps covers endpoints of the assets API seems to have changed a while ago (since May 26th 2021 with https://github.com/ppy/scthumber/pull/1 ?), but ___most___\* graveyard beatmaps that were not updated since then still have the old format. This includes the `card.jpg`, `card@2x.jpg`, `list.jpg` and `list@2x.jpg` images. (cover.jpg and slimcover.jpg did not change) Graveyard maps are often used in tournaments, and the card/list image are used in osu-web, in lazer (?), and by third-parties applications like mappool sheets. Example of user favourite section where the graveyard maps have a blurrier look due to this : ![image](https://user-images.githubusercontent.com/10669609/172273470-10b05442-49c9-451e-aeb3-0622c5c53271.png) Example graveyard map with the new format and image URLs : https://osu.ppy.sh/beatmapsets/1558526 https://assets.ppy.sh/beatmaps/1558526/covers/card.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/list.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/card@2x.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/list@2x.jpg?1636360215 (ranked and loved beatmaps seem to all return this format) Example graveyard map with the old format and image URLs : https://osu.ppy.sh/beatmapsets/1235496 https://assets.ppy.sh/beatmaps/1235496/covers/card.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/list.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/card@2x.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/list@2x.jpg?1599939054 *: I said most because there seems to be a few exceptions like https://osu.ppy.sh/beatmapsets/1306572 (last updated 22 Nov 2020, yet has the new format) Side-Note : Removing timestamp query parameter at the end of the URL seems to give the same result. What's the use of this timestamp query argument ? Retrieving old covers ?
1.0
Beatmap covers API endpoints return images of the old format for most non-recent graveyard maps - The size of images returned by the beatmaps covers endpoints of the assets API seems to have changed a while ago (since May 26th 2021 with https://github.com/ppy/scthumber/pull/1 ?), but ___most___\* graveyard beatmaps that were not updated since then still have the old format. This includes the `card.jpg`, `card@2x.jpg`, `list.jpg` and `list@2x.jpg` images. (cover.jpg and slimcover.jpg did not change) Graveyard maps are often used in tournaments, and the card/list image are used in osu-web, in lazer (?), and by third-parties applications like mappool sheets. Example of user favourite section where the graveyard maps have a blurrier look due to this : ![image](https://user-images.githubusercontent.com/10669609/172273470-10b05442-49c9-451e-aeb3-0622c5c53271.png) Example graveyard map with the new format and image URLs : https://osu.ppy.sh/beatmapsets/1558526 https://assets.ppy.sh/beatmaps/1558526/covers/card.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/list.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/card@2x.jpg?1636360215 https://assets.ppy.sh/beatmaps/1558526/covers/list@2x.jpg?1636360215 (ranked and loved beatmaps seem to all return this format) Example graveyard map with the old format and image URLs : https://osu.ppy.sh/beatmapsets/1235496 https://assets.ppy.sh/beatmaps/1235496/covers/card.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/list.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/card@2x.jpg?1599939054 https://assets.ppy.sh/beatmaps/1235496/covers/list@2x.jpg?1599939054 *: I said most because there seems to be a few exceptions like https://osu.ppy.sh/beatmapsets/1306572 (last updated 22 Nov 2020, yet has the new format) Side-Note : Removing timestamp query parameter at the end of the URL seems to give the same result. What's the use of this timestamp query argument ? Retrieving old covers ?
process
beatmap covers api endpoints return images of the old format for most non recent graveyard maps the size of images returned by the beatmaps covers endpoints of the assets api seems to have changed a while ago since may with but most graveyard beatmaps that were not updated since then still have the old format this includes the card jpg card jpg list jpg and list jpg images cover jpg and slimcover jpg did not change graveyard maps are often used in tournaments and the card list image are used in osu web in lazer and by third parties applications like mappool sheets example of user favourite section where the graveyard maps have a blurrier look due to this example graveyard map with the new format and image urls ranked and loved beatmaps seem to all return this format example graveyard map with the old format and image urls i said most because there seems to be a few exceptions like last updated nov yet has the new format side note removing timestamp query parameter at the end of the url seems to give the same result what s the use of this timestamp query argument retrieving old covers
1
67,624
17,023,035,571
IssuesEvent
2021-07-03 00:07:03
python-pillow/Pillow
https://api.github.com/repos/python-pillow/Pillow
reopened
Pillow built with libraqm insists it has no libraqm
Build Windows
<!-- Thank you for reporting an issue. Follow these guidelines to ensure your issue is handled properly. If you have a ... 1. General question: consider asking the question on Stack Overflow with the python-imaging-library tag: * https://stackoverflow.com/questions/tagged/python-imaging-library Do not ask a question in both places. If you think you have found a bug or have an unexplained exception then file a bug report here. 2. Bug report: include a self-contained, copy-pastable example that generates the issue if possible. Be concise with code posted. Guidelines on how to provide a good bug report: * https://stackoverflow.com/help/mcve Bug reports which follow these guidelines are easier to diagnose, and are often handled much more quickly. 3. Feature request: do a quick search of existing issues to make sure this has not been asked before. We know asking good questions takes effort, and we appreciate your time. Thank you. --> ### What did you do? Built pillow from source with all dependencies and installed it using `python setup.py install`. Running selftest.py works, and seems to corroborate that raqm is [indeed installed](https://justpaste.it/9e93v). ### What did you expect to happen? I expected to be able to set text direction ### What actually happened? I got the following exception: ``` File "C:\venv\lib\site-packages\pillow-7.2.0-py3.7-win-amd64.egg\PIL\ImageFont.py", line 262, in getsize KeyError: 'setting text direction, language or font features is not supported without libraqm' ``` ### What are your OS, Python and Pillow versions? * OS: Windows 10 * Python: 3.7.7 * Pillow: 7.2.0 <!-- Please include **code** that reproduces the issue and whenever possible, an **image** that demonstrates the issue. Please upload images to GitHub, not to third-party file hosting sites. If necessary, add the image to a zip or tar archive. The best reproductions are self-contained scripts with minimal dependencies. If you are using a framework such as Plone, Django, or Buildout, try to replicate the issue just using Pillow. --> ```python from PIL import ImageFont font = ImageFont.truetype(r'C:\Windows\Fonts\Arial.ttf', 120) size = font.getsize("hello world", 'rtl') ```
1.0
Pillow built with libraqm insists it has no libraqm - <!-- Thank you for reporting an issue. Follow these guidelines to ensure your issue is handled properly. If you have a ... 1. General question: consider asking the question on Stack Overflow with the python-imaging-library tag: * https://stackoverflow.com/questions/tagged/python-imaging-library Do not ask a question in both places. If you think you have found a bug or have an unexplained exception then file a bug report here. 2. Bug report: include a self-contained, copy-pastable example that generates the issue if possible. Be concise with code posted. Guidelines on how to provide a good bug report: * https://stackoverflow.com/help/mcve Bug reports which follow these guidelines are easier to diagnose, and are often handled much more quickly. 3. Feature request: do a quick search of existing issues to make sure this has not been asked before. We know asking good questions takes effort, and we appreciate your time. Thank you. --> ### What did you do? Built pillow from source with all dependencies and installed it using `python setup.py install`. Running selftest.py works, and seems to corroborate that raqm is [indeed installed](https://justpaste.it/9e93v). ### What did you expect to happen? I expected to be able to set text direction ### What actually happened? I got the following exception: ``` File "C:\venv\lib\site-packages\pillow-7.2.0-py3.7-win-amd64.egg\PIL\ImageFont.py", line 262, in getsize KeyError: 'setting text direction, language or font features is not supported without libraqm' ``` ### What are your OS, Python and Pillow versions? * OS: Windows 10 * Python: 3.7.7 * Pillow: 7.2.0 <!-- Please include **code** that reproduces the issue and whenever possible, an **image** that demonstrates the issue. Please upload images to GitHub, not to third-party file hosting sites. If necessary, add the image to a zip or tar archive. The best reproductions are self-contained scripts with minimal dependencies. If you are using a framework such as Plone, Django, or Buildout, try to replicate the issue just using Pillow. --> ```python from PIL import ImageFont font = ImageFont.truetype(r'C:\Windows\Fonts\Arial.ttf', 120) size = font.getsize("hello world", 'rtl') ```
non_process
pillow built with libraqm insists it has no libraqm thank you for reporting an issue follow these guidelines to ensure your issue is handled properly if you have a general question consider asking the question on stack overflow with the python imaging library tag do not ask a question in both places if you think you have found a bug or have an unexplained exception then file a bug report here bug report include a self contained copy pastable example that generates the issue if possible be concise with code posted guidelines on how to provide a good bug report bug reports which follow these guidelines are easier to diagnose and are often handled much more quickly feature request do a quick search of existing issues to make sure this has not been asked before we know asking good questions takes effort and we appreciate your time thank you what did you do built pillow from source with all dependencies and installed it using python setup py install running selftest py works and seems to corroborate that raqm is what did you expect to happen i expected to be able to set text direction what actually happened i got the following exception file c venv lib site packages pillow win egg pil imagefont py line in getsize keyerror setting text direction language or font features is not supported without libraqm what are your os python and pillow versions os windows python pillow please include code that reproduces the issue and whenever possible an image that demonstrates the issue please upload images to github not to third party file hosting sites if necessary add the image to a zip or tar archive the best reproductions are self contained scripts with minimal dependencies if you are using a framework such as plone django or buildout try to replicate the issue just using pillow python from pil import imagefont font imagefont truetype r c windows fonts arial ttf size font getsize hello world rtl
0
11,641
14,496,692,839
IssuesEvent
2020-12-11 13:10:54
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
reopened
Support Custom Logs schema updates
epic p1 team:data processing
### Description Users should be able to update schemas that they have already created in order to: 1. Add a new column 2. Add an indicator to an existing column 3. Investigate if possible to remove existing columns ### RFC TBD ### Designs TBD ### Acceptance Criteria - Users can add a column to an existing schema. The new column should be queryable through the data lake and available to the streaming rules - Users can add an indicator to an existing column. - In case an update is not possible, the UI should notify the users
1.0
Support Custom Logs schema updates - ### Description Users should be able to update schemas that they have already created in order to: 1. Add a new column 2. Add an indicator to an existing column 3. Investigate if possible to remove existing columns ### RFC TBD ### Designs TBD ### Acceptance Criteria - Users can add a column to an existing schema. The new column should be queryable through the data lake and available to the streaming rules - Users can add an indicator to an existing column. - In case an update is not possible, the UI should notify the users
process
support custom logs schema updates description users should be able to update schemas that they have already created in order to add a new column add an indicator to an existing column investigate if possible to remove existing columns rfc tbd designs tbd acceptance criteria users can add a column to an existing schema the new column should be queryable through the data lake and available to the streaming rules users can add an indicator to an existing column in case an update is not possible the ui should notify the users
1
3,517
6,562,079,186
IssuesEvent
2017-09-07 15:21:28
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
opened
allow preprocessor to accept text input instead of mod file
enhancement preprocessor
for GUI interface, create a preprocessor argument that would be the `.mod` file in string format. This way the GUI would not have to make a `.mod` file every time the user changed something
1.0
allow preprocessor to accept text input instead of mod file - for GUI interface, create a preprocessor argument that would be the `.mod` file in string format. This way the GUI would not have to make a `.mod` file every time the user changed something
process
allow preprocessor to accept text input instead of mod file for gui interface create a preprocessor argument that would be the mod file in string format this way the gui would not have to make a mod file every time the user changed something
1
7,193
10,331,810,353
IssuesEvent
2019-09-02 19:58:13
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
Facebook Domain False Positive
whitelisting process
*@hyperxpro commented on Jan 13, 2019, 8:31 PM UTC:* False Positive Facebook Domains Just To Name A Few: ## 3-edge-chat.facebook.com ## 3-edge-chat.facebook.com ## 5-edge-chat.facebook.com ## static.xx.fbcdn.net *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#503](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/503).*
1.0
Facebook Domain False Positive - *@hyperxpro commented on Jan 13, 2019, 8:31 PM UTC:* False Positive Facebook Domains Just To Name A Few: ## 3-edge-chat.facebook.com ## 3-edge-chat.facebook.com ## 5-edge-chat.facebook.com ## static.xx.fbcdn.net *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#503](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/503).*
process
facebook domain false positive hyperxpro commented on jan pm utc false positive facebook domains just to name a few edge chat facebook com edge chat facebook com edge chat facebook com static xx fbcdn net this issue was moved by from
1
87
2,534,269,573
IssuesEvent
2015-01-24 19:46:45
ufvivotech/ufDataQualityImprovement
https://api.github.com/repos/ufvivotech/ufDataQualityImprovement
closed
Create role management ingest
Ontology Process Software Dev
Role management ingest would allow for associating roles to people. Often roles involves associating people to organizations -- CTSI investigator status, committee memberships, etc. We should have a simple reproducible, five case logic method for managing roles. Software has been written. Needs testing and deployment and process.
1.0
Create role management ingest - Role management ingest would allow for associating roles to people. Often roles involves associating people to organizations -- CTSI investigator status, committee memberships, etc. We should have a simple reproducible, five case logic method for managing roles. Software has been written. Needs testing and deployment and process.
process
create role management ingest role management ingest would allow for associating roles to people often roles involves associating people to organizations ctsi investigator status committee memberships etc we should have a simple reproducible five case logic method for managing roles software has been written needs testing and deployment and process
1
179,227
30,199,770,056
IssuesEvent
2023-07-05 03:49:30
microsoft/devhome
https://api.github.com/repos/microsoft/devhome
reopened
Executing Tasks should provide more information.
Issue-Bug Area-Machine-Config Resolution-By-Design Area-Machine-Config-Loading
### Dev Home version _No response_ ### Windows build number _No response_ ### Other software _No response_ ### Steps to reproduce the bug TODO here: DevHome.SetupFlow.ISetupTasks.cs ### Expected result _No response_ ### Actual result _No response_ ### Included System Information _No response_ ### Included Extensions Information _No response_
1.0
Executing Tasks should provide more information. - ### Dev Home version _No response_ ### Windows build number _No response_ ### Other software _No response_ ### Steps to reproduce the bug TODO here: DevHome.SetupFlow.ISetupTasks.cs ### Expected result _No response_ ### Actual result _No response_ ### Included System Information _No response_ ### Included Extensions Information _No response_
non_process
executing tasks should provide more information dev home version no response windows build number no response other software no response steps to reproduce the bug todo here devhome setupflow isetuptasks cs expected result no response actual result no response included system information no response included extensions information no response
0
224,867
17,202,879,640
IssuesEvent
2021-07-17 16:19:06
oasis-tcs/csaf
https://api.github.com/repos/oasis-tcs/csaf
opened
Sort enumeration value listings
csaf 2.0 documentation editorial enhancement
Some enumeration value listings in the schema (and of course) prose still do not follow a lexical ordering. I propose we sort the enum values to make clear, that the order of those values has no further semantics. Later additions of the spec should be empowered to add (possible) or remove (improbable) such catalog values and then simply resort the list. Any implementation is well advised to not rely on some numerical ordering of an underlying representation like the index number of the JSON array we use to denote the values in the schema.
1.0
Sort enumeration value listings - Some enumeration value listings in the schema (and of course) prose still do not follow a lexical ordering. I propose we sort the enum values to make clear, that the order of those values has no further semantics. Later additions of the spec should be empowered to add (possible) or remove (improbable) such catalog values and then simply resort the list. Any implementation is well advised to not rely on some numerical ordering of an underlying representation like the index number of the JSON array we use to denote the values in the schema.
non_process
sort enumeration value listings some enumeration value listings in the schema and of course prose still do not follow a lexical ordering i propose we sort the enum values to make clear that the order of those values has no further semantics later additions of the spec should be empowered to add possible or remove improbable such catalog values and then simply resort the list any implementation is well advised to not rely on some numerical ordering of an underlying representation like the index number of the json array we use to denote the values in the schema
0
1,349
3,908,213,174
IssuesEvent
2016-04-19 15:13:59
BEP-store/project-plan
https://api.github.com/repos/BEP-store/project-plan
closed
Project planning
Presentation Process Product Report Research
An outline of the milestones and due dates of the project should be included in the project plan.
1.0
Project planning - An outline of the milestones and due dates of the project should be included in the project plan.
process
project planning an outline of the milestones and due dates of the project should be included in the project plan
1
18,049
24,058,243,956
IssuesEvent
2022-09-16 19:07:12
GoogleCloudPlatform/terraform-mean-cloudrun-mongodb
https://api.github.com/repos/GoogleCloudPlatform/terraform-mean-cloudrun-mongodb
closed
Prepare initial MEAN stack reference app to use for testing
process
https://github.com/mongodb-developer/mean-stack-example Link to public container images: Server: [gcr.io/abis-345004/mean-stack-example](http://gcr.io/abis-345004/mean-stack-example) Client: [gcr.io/abis-345004/mean-stack-example_client](http://gcr.io/abis-345004/mean-stack-example_client)
1.0
Prepare initial MEAN stack reference app to use for testing - https://github.com/mongodb-developer/mean-stack-example Link to public container images: Server: [gcr.io/abis-345004/mean-stack-example](http://gcr.io/abis-345004/mean-stack-example) Client: [gcr.io/abis-345004/mean-stack-example_client](http://gcr.io/abis-345004/mean-stack-example_client)
process
prepare initial mean stack reference app to use for testing link to public container images server client
1
236,443
7,749,254,134
IssuesEvent
2018-05-30 10:50:49
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0003632: Method not found: Crm.getSettings
Admin bug high priority
**Reported by pschuele on 3 Jan 2011 16:42** **Version:** Neele (2011-01) Beta 3 Method not found: Crm.getSettings in Admin/Applications/Crm/Settings
1.0
0003632: Method not found: Crm.getSettings - **Reported by pschuele on 3 Jan 2011 16:42** **Version:** Neele (2011-01) Beta 3 Method not found: Crm.getSettings in Admin/Applications/Crm/Settings
non_process
method not found crm getsettings reported by pschuele on jan version neele beta method not found crm getsettings in admin applications crm settings
0
15,743
19,910,546,736
IssuesEvent
2022-01-25 16:44:37
input-output-hk/high-assurance-legacy
https://api.github.com/repos/input-output-hk/high-assurance-legacy
closed
Formally prove `distributor` core lemma
type: enhancement language: isabelle topic: process calculus
Our goal is to formally prove the `distributor` core lemma described in #36. An informal proof of this lemma is given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
1.0
Formally prove `distributor` core lemma - Our goal is to formally prove the `distributor` core lemma described in #36. An informal proof of this lemma is given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
process
formally prove distributor core lemma our goal is to formally prove the distributor core lemma described in an informal proof of this lemma is given in
1
90,687
15,856,233,871
IssuesEvent
2021-04-08 01:51:23
heholek/charted
https://api.github.com/repos/heholek/charted
opened
CVE-2015-8315 (High) detected in ms-0.6.2.tgz, ms-0.7.0.tgz
security vulnerability
## CVE-2015-8315 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ms-0.6.2.tgz</b>, <b>ms-0.7.0.tgz</b></p></summary> <p> <details><summary><b>ms-0.6.2.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.6.2.tgz">https://registry.npmjs.org/ms/-/ms-0.6.2.tgz</a></p> <p>Path to dependency file: charted/package.json</p> <p>Path to vulnerable library: charted/node_modules/send/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - express-4.10.8.tgz (Root Library) - send-0.10.1.tgz - :x: **ms-0.6.2.tgz** (Vulnerable Library) </details> <details><summary><b>ms-0.7.0.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.0.tgz">https://registry.npmjs.org/ms/-/ms-0.7.0.tgz</a></p> <p>Path to dependency file: charted/package.json</p> <p>Path to vulnerable library: charted/node_modules/send/node_modules/debug/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - express-4.10.8.tgz (Root Library) - debug-2.1.3.tgz - :x: **ms-0.7.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ms package before 0.7.1 for Node.js allows attackers to cause a denial of service (CPU consumption) via a long version string, aka a "regular expression denial of service (ReDoS)." <p>Publish Date: 2017-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8315>CVE-2015-8315</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8315">https://nvd.nist.gov/vuln/detail/CVE-2015-8315</a></p> <p>Release Date: 2017-01-23</p> <p>Fix Resolution: 0.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-8315 (High) detected in ms-0.6.2.tgz, ms-0.7.0.tgz - ## CVE-2015-8315 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ms-0.6.2.tgz</b>, <b>ms-0.7.0.tgz</b></p></summary> <p> <details><summary><b>ms-0.6.2.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.6.2.tgz">https://registry.npmjs.org/ms/-/ms-0.6.2.tgz</a></p> <p>Path to dependency file: charted/package.json</p> <p>Path to vulnerable library: charted/node_modules/send/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - express-4.10.8.tgz (Root Library) - send-0.10.1.tgz - :x: **ms-0.6.2.tgz** (Vulnerable Library) </details> <details><summary><b>ms-0.7.0.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.0.tgz">https://registry.npmjs.org/ms/-/ms-0.7.0.tgz</a></p> <p>Path to dependency file: charted/package.json</p> <p>Path to vulnerable library: charted/node_modules/send/node_modules/debug/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - express-4.10.8.tgz (Root Library) - debug-2.1.3.tgz - :x: **ms-0.7.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ms package before 0.7.1 for Node.js allows attackers to cause a denial of service (CPU consumption) via a long version string, aka a "regular expression denial of service (ReDoS)." <p>Publish Date: 2017-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8315>CVE-2015-8315</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8315">https://nvd.nist.gov/vuln/detail/CVE-2015-8315</a></p> <p>Release Date: 2017-01-23</p> <p>Fix Resolution: 0.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in ms tgz ms tgz cve high severity vulnerability vulnerable libraries ms tgz ms tgz ms tgz tiny ms conversion utility library home page a href path to dependency file charted package json path to vulnerable library charted node modules send node modules ms package json dependency hierarchy express tgz root library send tgz x ms tgz vulnerable library ms tgz tiny ms conversion utility library home page a href path to dependency file charted package json path to vulnerable library charted node modules send node modules debug node modules ms package json dependency hierarchy express tgz root library debug tgz x ms tgz vulnerable library vulnerability details the ms package before for node js allows attackers to cause a denial of service cpu consumption via a long version string aka a regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
15,249
26,778,020,768
IssuesEvent
2023-01-31 18:41:38
NASA-PDS/registry
https://api.github.com/repos/NASA-PDS/registry
opened
As a user, I want to track the deep archive status for a product
needs:triage requirement
### Checked for duplicates Yes - I've already checked ### 🧑‍🔬 User Persona(s) Node Operator NSSDCA Operator ### 💪 Motivation ...so that I can know if a product has been successfully ingested by the NSSDCA. ### 📖 Additional Details _No response_ ### Acceptance Criteria **Given** a product in the registry **When I perform** a query of that products tracking metadata information **Then I expect** to know what the status is for the delivery and ingestion of the product into the NSSDCA. ### ⚙️ Engineering Details Statuses: TBD. should talk to NSSDCA for what their statuses are
1.0
As a user, I want to track the deep archive status for a product - ### Checked for duplicates Yes - I've already checked ### 🧑‍🔬 User Persona(s) Node Operator NSSDCA Operator ### 💪 Motivation ...so that I can know if a product has been successfully ingested by the NSSDCA. ### 📖 Additional Details _No response_ ### Acceptance Criteria **Given** a product in the registry **When I perform** a query of that products tracking metadata information **Then I expect** to know what the status is for the delivery and ingestion of the product into the NSSDCA. ### ⚙️ Engineering Details Statuses: TBD. should talk to NSSDCA for what their statuses are
non_process
as a user i want to track the deep archive status for a product checked for duplicates yes i ve already checked 🧑‍🔬 user persona s node operator nssdca operator 💪 motivation so that i can know if a product has been successfully ingested by the nssdca 📖 additional details no response acceptance criteria given a product in the registry when i perform a query of that products tracking metadata information then i expect to know what the status is for the delivery and ingestion of the product into the nssdca ⚙️ engineering details statuses tbd should talk to nssdca for what their statuses are
0
5,316
3,917,208,485
IssuesEvent
2016-04-21 07:12:33
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_ARM Assembly Language Programming - 2_Back to experiment_smk
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description : In the "ARM Assembly Language Programming - 2 " experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user. Actual Result : In the "ARM Assembly Language Programming - 2 " experiment,the back to experiments link is not present in the page. Environment : OS: Windows 7, Linux Browsers: Firefox,Chrome Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/omputer-organization-iiith/blob/master/test-cases/integration_test-cases/ARM%20Assembly%20Language%20Programming%20-%202/ARM%20Assembly%20Language%20Programming%20-%202_17_Back%20to%20experiment_smk.org
True
QA_ARM Assembly Language Programming - 2_Back to experiment_smk - Defect Description : In the "ARM Assembly Language Programming - 2 " experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user. Actual Result : In the "ARM Assembly Language Programming - 2 " experiment,the back to experiments link is not present in the page. Environment : OS: Windows 7, Linux Browsers: Firefox,Chrome Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/omputer-organization-iiith/blob/master/test-cases/integration_test-cases/ARM%20Assembly%20Language%20Programming%20-%202/ARM%20Assembly%20Language%20Programming%20-%202_17_Back%20to%20experiment_smk.org
non_process
qa arm assembly language programming back to experiment smk defect description in the arm assembly language programming experiment the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in order to view the list of experiments by the user actual result in the arm assembly language programming experiment the back to experiments link is not present in the page environment os windows linux browsers firefox chrome bandwidth hardware configuration processor test step link
0
17,126
22,647,262,053
IssuesEvent
2022-07-01 09:55:37
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
opened
Memory bursts occurred on Radiocontrast API
issue-processing-state-01
**Describe the bug** Memory bursts may occur when using the Radiocontrast API. For example, when taking [this Dex](https://github.com/quark-engine/quark-engine/files/9027918/flubot_payload.zip) as input and generating rules from the method `Lcom/iqiyi/i18n/p4c4a83bb;->onAccessibilityEvent(Landroid/view/accessibility/AccessibilityEvent;)V`, Quark v22.6.1 requires more than **64 GB of memory** to finish. However, compared to the size of the APK (2.1 MB), such high memory requirements may decrease the usability of the API. Thus, we may need more improvement on its memory usage. **Screenshot** ![](https://i.imgur.com/8gfqbdh.png) **To Reproduce** Start the rule generation by running the following script. ```python #!/usr/bin/python3 from quark.radiocontrast import RadioContrast APK_PATH = "/flubot_payload.dex" TARGET_METHOD = "Lcom/iqiyi/i18n/p4c4a83bb;->onAccessibilityEvent(Landroid/view/accessibility/AccessibilityEvent;)V" GENERATED_RULE_DIR = "./Generated" radiocontrast = RadioContrast( APK_PATH, TARGET_METHOD, GENERATED_RULE_DIR ) radiocontrast.rule_generate() ``` **Environment** + Ubuntu 18.04 + Python 3.8.10 + Quark-Engine v22.6.1 **Possible related issues** + #299 + #130
1.0
Memory bursts occurred on Radiocontrast API - **Describe the bug** Memory bursts may occur when using the Radiocontrast API. For example, when taking [this Dex](https://github.com/quark-engine/quark-engine/files/9027918/flubot_payload.zip) as input and generating rules from the method `Lcom/iqiyi/i18n/p4c4a83bb;->onAccessibilityEvent(Landroid/view/accessibility/AccessibilityEvent;)V`, Quark v22.6.1 requires more than **64 GB of memory** to finish. However, compared to the size of the APK (2.1 MB), such high memory requirements may decrease the usability of the API. Thus, we may need more improvement on its memory usage. **Screenshot** ![](https://i.imgur.com/8gfqbdh.png) **To Reproduce** Start the rule generation by running the following script. ```python #!/usr/bin/python3 from quark.radiocontrast import RadioContrast APK_PATH = "/flubot_payload.dex" TARGET_METHOD = "Lcom/iqiyi/i18n/p4c4a83bb;->onAccessibilityEvent(Landroid/view/accessibility/AccessibilityEvent;)V" GENERATED_RULE_DIR = "./Generated" radiocontrast = RadioContrast( APK_PATH, TARGET_METHOD, GENERATED_RULE_DIR ) radiocontrast.rule_generate() ``` **Environment** + Ubuntu 18.04 + Python 3.8.10 + Quark-Engine v22.6.1 **Possible related issues** + #299 + #130
process
memory bursts occurred on radiocontrast api describe the bug memory bursts may occur when using the radiocontrast api for example when taking as input and generating rules from the method lcom iqiyi onaccessibilityevent landroid view accessibility accessibilityevent v quark requires more than gb of memory to finish however compared to the size of the apk mb such high memory requirements may decrease the usability of the api thus we may need more improvement on its memory usage screenshot to reproduce start the rule generation by running the following script python usr bin from quark radiocontrast import radiocontrast apk path flubot payload dex target method lcom iqiyi onaccessibilityevent landroid view accessibility accessibilityevent v generated rule dir generated radiocontrast radiocontrast apk path target method generated rule dir radiocontrast rule generate environment ubuntu python quark engine possible related issues
1
1,563
4,162,614,943
IssuesEvent
2016-06-17 21:06:50
nodejs/node
https://api.github.com/repos/nodejs/node
opened
process, doc: discourage use of process.exit() for control flow in docs
doc process
Userland library authors have a pattern for CLIs that has lead to various issues during the introduction of v6. We had a heated discussion that I would summarize as: the below pattern should have been used in like this, since it actually never ensured to deliver what it promised; but core has a historic responsibility of not breaking such widely adopted patterns. ```js process.on('exit', () => { // do some post action here later }) function doSomething() { for (var i = 0; i < 1000; i++) { process.stdout.write('some result' + i + '\n') } // decide that the execution of the CLI should end here process.exit() } ``` The problem with this that stdout doesn't get flushed on `process.exit()`, resulting in not all all 1000 calls being printed. This became apparent in v6. Eventually this calls exit(3). In any good c++ practice exit(3) is discouraged, since functions scopes are not guaranteed to unwind properly. Imo, authors should at be just return from functions, from top scope, or use proper event emitters. Someone could open a doc PR. Ref: https://github.com/nodejs/node/issues/6980, https://github.com/nodejs/node/issues/6456 cc @Fishrock123
1.0
process, doc: discourage use of process.exit() for control flow in docs - Userland library authors have a pattern for CLIs that has lead to various issues during the introduction of v6. We had a heated discussion that I would summarize as: the below pattern should have been used in like this, since it actually never ensured to deliver what it promised; but core has a historic responsibility of not breaking such widely adopted patterns. ```js process.on('exit', () => { // do some post action here later }) function doSomething() { for (var i = 0; i < 1000; i++) { process.stdout.write('some result' + i + '\n') } // decide that the execution of the CLI should end here process.exit() } ``` The problem with this that stdout doesn't get flushed on `process.exit()`, resulting in not all all 1000 calls being printed. This became apparent in v6. Eventually this calls exit(3). In any good c++ practice exit(3) is discouraged, since functions scopes are not guaranteed to unwind properly. Imo, authors should at be just return from functions, from top scope, or use proper event emitters. Someone could open a doc PR. Ref: https://github.com/nodejs/node/issues/6980, https://github.com/nodejs/node/issues/6456 cc @Fishrock123
process
process doc discourage use of process exit for control flow in docs userland library authors have a pattern for clis that has lead to various issues during the introduction of we had a heated discussion that i would summarize as the below pattern should have been used in like this since it actually never ensured to deliver what it promised but core has a historic responsibility of not breaking such widely adopted patterns js process on exit do some post action here later function dosomething for var i i i process stdout write some result i n decide that the execution of the cli should end here process exit the problem with this that stdout doesn t get flushed on process exit resulting in not all all calls being printed this became apparent in eventually this calls exit in any good c practice exit is discouraged since functions scopes are not guaranteed to unwind properly imo authors should at be just return from functions from top scope or use proper event emitters someone could open a doc pr ref cc
1
112,388
14,244,128,066
IssuesEvent
2020-11-19 06:15:34
urbit/landscape
https://api.github.com/repos/urbit/landscape
closed
landscape: setting root profile should set profile everywhere
design
"Contacts per group" doesn't really make sense, and should be architected to be more sane, allowing you to store one profile per ship you interact with.
1.0
landscape: setting root profile should set profile everywhere - "Contacts per group" doesn't really make sense, and should be architected to be more sane, allowing you to store one profile per ship you interact with.
non_process
landscape setting root profile should set profile everywhere contacts per group doesn t really make sense and should be architected to be more sane allowing you to store one profile per ship you interact with
0
51,346
7,696,395,696
IssuesEvent
2018-05-18 15:11:39
Adobe-Marketing-Cloud/aem-core-wcm-components
https://api.github.com/repos/Adobe-Marketing-Cloud/aem-core-wcm-components
closed
List Component Pagination
documentation question
The documentation says the List Compnent provides: Ordering, pagination and limit. I see both Ordering and Limit, but I do not see a way to set Pagination. Is the Pagination feature implemented?
1.0
List Component Pagination - The documentation says the List Compnent provides: Ordering, pagination and limit. I see both Ordering and Limit, but I do not see a way to set Pagination. Is the Pagination feature implemented?
non_process
list component pagination the documentation says the list compnent provides ordering pagination and limit i see both ordering and limit but i do not see a way to set pagination is the pagination feature implemented
0
411,778
27,830,558,564
IssuesEvent
2023-03-20 04:10:42
1C-Company/v8-code-style
https://api.github.com/repos/1C-Company/v8-code-style
opened
Добавить описание doc-comment-return-section-type.md
documentation
## Раздел документации или код проверки <!-- Путь к разделу документации: --> `doc-comment-return-section-type.md` ## Что необходимо улучшить <!-- Кратко опишите, что нужно улучшить, исправить. --> Добавить описание
1.0
Добавить описание doc-comment-return-section-type.md - ## Раздел документации или код проверки <!-- Путь к разделу документации: --> `doc-comment-return-section-type.md` ## Что необходимо улучшить <!-- Кратко опишите, что нужно улучшить, исправить. --> Добавить описание
non_process
добавить описание doc comment return section type md раздел документации или код проверки doc comment return section type md что необходимо улучшить добавить описание
0
173,948
13,450,296,926
IssuesEvent
2020-09-08 18:17:48
certbot/certbot
https://api.github.com/repos/certbot/certbot
closed
Include server logs in certbot-ci output
area: boulder area: error handling area: testing needs-update
We've been seeing a lot of spurious test failures in our integration tests in Travis where we time out waiting for boulder to start. I initially hoped it was just we're not waiting long enough, but we increased the timeout to 4 minutes and it still regularly fails. I think the best way to start fixing this is to include server logs in the Travis output at least when the server fails to start so we can get more information about what's going wrong. I think we should make sure that the client and server output is never intermixed. This could be done by storing the server's output in memory or in a logfile and at the end of execution (probably only if tests fail?) dump that output to stdout either in the `certbot-ci` code or print the logfile in our CI config.
1.0
Include server logs in certbot-ci output - We've been seeing a lot of spurious test failures in our integration tests in Travis where we time out waiting for boulder to start. I initially hoped it was just we're not waiting long enough, but we increased the timeout to 4 minutes and it still regularly fails. I think the best way to start fixing this is to include server logs in the Travis output at least when the server fails to start so we can get more information about what's going wrong. I think we should make sure that the client and server output is never intermixed. This could be done by storing the server's output in memory or in a logfile and at the end of execution (probably only if tests fail?) dump that output to stdout either in the `certbot-ci` code or print the logfile in our CI config.
non_process
include server logs in certbot ci output we ve been seeing a lot of spurious test failures in our integration tests in travis where we time out waiting for boulder to start i initially hoped it was just we re not waiting long enough but we increased the timeout to minutes and it still regularly fails i think the best way to start fixing this is to include server logs in the travis output at least when the server fails to start so we can get more information about what s going wrong i think we should make sure that the client and server output is never intermixed this could be done by storing the server s output in memory or in a logfile and at the end of execution probably only if tests fail dump that output to stdout either in the certbot ci code or print the logfile in our ci config
0
14,524
17,620,027,942
IssuesEvent
2021-08-18 14:20:14
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Calculate Binned Weather Results in Cooling Tower Basin calc
enhancement Process Cooling
For each bin set, run calculation method with input parameter = parameter * binned hours. May contain multiple parameters
1.0
Calculate Binned Weather Results in Cooling Tower Basin calc - For each bin set, run calculation method with input parameter = parameter * binned hours. May contain multiple parameters
process
calculate binned weather results in cooling tower basin calc for each bin set run calculation method with input parameter parameter binned hours may contain multiple parameters
1
4,594
7,432,197,694
IssuesEvent
2018-03-25 22:10:30
nyu-software-engineering/mood-music-player
https://api.github.com/repos/nyu-software-engineering/mood-music-player
closed
As a user, I want to be able to connect my Spotify account.
2 - in process user story
- [x] make profile page UI with "connect to Spotify" button - [x] redirect to Spotify to get authorization when you click the button - [x] redirect back to browse after getting auth token
1.0
As a user, I want to be able to connect my Spotify account. - - [x] make profile page UI with "connect to Spotify" button - [x] redirect to Spotify to get authorization when you click the button - [x] redirect back to browse after getting auth token
process
as a user i want to be able to connect my spotify account make profile page ui with connect to spotify button redirect to spotify to get authorization when you click the button redirect back to browse after getting auth token
1
204,352
15,439,111,616
IssuesEvent
2021-03-07 22:57:44
mainlyetcetera/quick-quotes
https://api.github.com/repos/mainlyetcetera/quick-quotes
closed
Implement E2E Testing
enhancement testing
**Description** > What does this addition do? What problems does it solve? - ensure solid app with diff data sets - test for both happy and sad paths for user **Planning** > Is there a specific plan for how to implement this? What needs to change? - stub data - design happy and sad path tests
1.0
Implement E2E Testing - **Description** > What does this addition do? What problems does it solve? - ensure solid app with diff data sets - test for both happy and sad paths for user **Planning** > Is there a specific plan for how to implement this? What needs to change? - stub data - design happy and sad path tests
non_process
implement testing description what does this addition do what problems does it solve ensure solid app with diff data sets test for both happy and sad paths for user planning is there a specific plan for how to implement this what needs to change stub data design happy and sad path tests
0
20,784
27,521,407,730
IssuesEvent
2023-03-06 15:15:29
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
qgis_process algorithms create temporary file 'srs6.db' without cleaning
Processing Bug
### What is the bug or the crash? In the dev version, apparently all algorithms used with `qgis_process` result in a temporary file `srs6.db` which is not removed after finishing. The algorithms that have been tested and where this occurs (none were found where it doesn't happen): ``` native:buffer native:createconstantrasterlayer native:simplifygeometries native:splitvectorlayer native:extractbyattribute native:reprojectlayer native:mergevectorlayers native:printlayouttopdf qgis:relief ``` ### Steps to reproduce the issue ```bash $ mkdir ~/test $ cd ~/test $ pwd /home/floris/test $ $ find /tmp -type f -name srs* 2> /dev/null $ $ myJSON=$(</dev/stdin) { "area_units": "m2", "distance_units": "meters", "ellipsoid": "EPSG:7030", "inputs": { "FIELD": "NAME", "INPUT": "/home/floris/git_repositories2/QGIS/build-216fdff9/output/data/resources/data/world_map.gpkg|layername=countries", "OPERATOR": 0, "OUTPUT": "ogr:dbname='/home/floris/test/output_country.gpkg' table=\"output\" (geom)", "VALUE": "Norway" } } $ $ echo "$myJSON" | ~/git_repositories2/QGIS/build-216fdff9/output/bin/qgis_process run native:extractbyattribute - 1> /dev/null $ $ ls ~/test output_country.gpkg $ $ find /tmp -type f -mmin -1 2> /dev/null /tmp/srs6.db ``` ### Versions ```bash $ ~/git_repositories2/QGIS/build-216fdff9/output/bin/qgis_process --version QGIS 3.29.0-Master 'Master' (216fdff9) QGIS code revision 216fdff9 Qt version 5.15.3 Python version 3.10.6 GDAL/OGR version 3.6.2 PROJ version 9.1.1 EPSG Registry database version v10.076 (2022-08-31) GEOS version 3.11.1-CAPI-1.17.1 SQLite version 3.37.2 OS Linux Mint 21.1 ``` ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context This behaviour did not occur in QGIS 3.28.2, in which related issue #50623 had been solved. Also, file `srs6.db` is new compared to the temp files found in that issue.
1.0
qgis_process algorithms create temporary file 'srs6.db' without cleaning - ### What is the bug or the crash? In the dev version, apparently all algorithms used with `qgis_process` result in a temporary file `srs6.db` which is not removed after finishing. The algorithms that have been tested and where this occurs (none were found where it doesn't happen): ``` native:buffer native:createconstantrasterlayer native:simplifygeometries native:splitvectorlayer native:extractbyattribute native:reprojectlayer native:mergevectorlayers native:printlayouttopdf qgis:relief ``` ### Steps to reproduce the issue ```bash $ mkdir ~/test $ cd ~/test $ pwd /home/floris/test $ $ find /tmp -type f -name srs* 2> /dev/null $ $ myJSON=$(</dev/stdin) { "area_units": "m2", "distance_units": "meters", "ellipsoid": "EPSG:7030", "inputs": { "FIELD": "NAME", "INPUT": "/home/floris/git_repositories2/QGIS/build-216fdff9/output/data/resources/data/world_map.gpkg|layername=countries", "OPERATOR": 0, "OUTPUT": "ogr:dbname='/home/floris/test/output_country.gpkg' table=\"output\" (geom)", "VALUE": "Norway" } } $ $ echo "$myJSON" | ~/git_repositories2/QGIS/build-216fdff9/output/bin/qgis_process run native:extractbyattribute - 1> /dev/null $ $ ls ~/test output_country.gpkg $ $ find /tmp -type f -mmin -1 2> /dev/null /tmp/srs6.db ``` ### Versions ```bash $ ~/git_repositories2/QGIS/build-216fdff9/output/bin/qgis_process --version QGIS 3.29.0-Master 'Master' (216fdff9) QGIS code revision 216fdff9 Qt version 5.15.3 Python version 3.10.6 GDAL/OGR version 3.6.2 PROJ version 9.1.1 EPSG Registry database version v10.076 (2022-08-31) GEOS version 3.11.1-CAPI-1.17.1 SQLite version 3.37.2 OS Linux Mint 21.1 ``` ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context This behaviour did not occur in QGIS 3.28.2, in which related issue #50623 had been solved. Also, file `srs6.db` is new compared to the temp files found in that issue.
process
qgis process algorithms create temporary file db without cleaning what is the bug or the crash in the dev version apparently all algorithms used with qgis process result in a temporary file db which is not removed after finishing the algorithms that have been tested and where this occurs none were found where it doesn t happen native buffer native createconstantrasterlayer native simplifygeometries native splitvectorlayer native extractbyattribute native reprojectlayer native mergevectorlayers native printlayouttopdf qgis relief steps to reproduce the issue bash mkdir test cd test pwd home floris test find tmp type f name srs dev null myjson dev stdin area units distance units meters ellipsoid epsg inputs field name input home floris git qgis build output data resources data world map gpkg layername countries operator output ogr dbname home floris test output country gpkg table output geom value norway echo myjson git qgis build output bin qgis process run native extractbyattribute dev null ls test output country gpkg find tmp type f mmin dev null tmp db versions bash git qgis build output bin qgis process version qgis master master qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version os linux mint supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context this behaviour did not occur in qgis in which related issue had been solved also file db is new compared to the temp files found in that issue
1
15,790
19,981,749,424
IssuesEvent
2022-01-30 01:46:15
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Firefox crashes when test running on the specific page
TYPE: bug AREA: client BROWSER: Firefox FREQUENCY: level 1 SYSTEM: client side processing STATE: Stale
It looks like memory leak. See the attached screencast ### What is your web application and your TestCafe test code? `fixture `New Fixture` .page `https://www.w3schools.com/bootstrap/bootstrap_forms_inputs.asp`; test('New Test', async t => { await t .debug(); });` ### Steps to Reproduce: https://www.screencast.com/t/4q8bZink https://www.screencast.com/t/AJyZ89oFvL ### Your Environment details: * testcafe version: `testcafe@1.1.1-rc.3` * node.js version: `10.15.3` * command-line arguments: "testcafe firefox test.js" * browser name and version: Firefox 66 * platform and version: Windows 8.1
1.0
Firefox crashes when test running on the specific page - It looks like memory leak. See the attached screencast ### What is your web application and your TestCafe test code? `fixture `New Fixture` .page `https://www.w3schools.com/bootstrap/bootstrap_forms_inputs.asp`; test('New Test', async t => { await t .debug(); });` ### Steps to Reproduce: https://www.screencast.com/t/4q8bZink https://www.screencast.com/t/AJyZ89oFvL ### Your Environment details: * testcafe version: `testcafe@1.1.1-rc.3` * node.js version: `10.15.3` * command-line arguments: "testcafe firefox test.js" * browser name and version: Firefox 66 * platform and version: Windows 8.1
process
firefox crashes when test running on the specific page it looks like memory leak see the attached screencast what is your web application and your testcafe test code fixture new fixture page test new test async t await t debug steps to reproduce your environment details testcafe version testcafe rc node js version command line arguments testcafe firefox test js browser name and version firefox platform and version windows
1
508,314
14,698,435,798
IssuesEvent
2021-01-04 06:26:38
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
hole-io.com - Unable to play the game
browser-firefox engine-gecko priority-normal severity-critical
<!-- @browser: Firefox 84.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/61308 --> **URL**: https://hole-io.com/?utm_source=gameads&utm_campaign=608881115&utm_term=paper-io.com# **Browser / Version**: Firefox 84.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: page is loading a lot of time <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/e4268ba6-ffca-445f-85b6-b8e3a99eea43.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201106160425</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/93fd5053-323f-4c3b-b82d-40c2763dc804) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
hole-io.com - Unable to play the game - <!-- @browser: Firefox 84.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/61308 --> **URL**: https://hole-io.com/?utm_source=gameads&utm_campaign=608881115&utm_term=paper-io.com# **Browser / Version**: Firefox 84.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: page is loading a lot of time <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/e4268ba6-ffca-445f-85b6-b8e3a99eea43.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201106160425</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/93fd5053-323f-4c3b-b82d-40c2763dc804) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
hole io com unable to play the game url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce page is loading a lot of time view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
104,822
16,622,392,757
IssuesEvent
2021-06-03 04:20:34
gms-ws-sandbox/NodeGoat
https://api.github.com/repos/gms-ws-sandbox/NodeGoat
opened
CVE-2018-1000620 (High) detected in cryptiles-0.2.2.tgz, cryptiles-2.0.5.tgz
security vulnerability
## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>cryptiles-0.2.2.tgz</b>, <b>cryptiles-2.0.5.tgz</b></p></summary> <p> <details><summary><b>cryptiles-0.2.2.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-0.2.2.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-0.2.2.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/zaproxy/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - hawk-1.0.0.tgz - :x: **cryptiles-0.2.2.tgz** (Vulnerable Library) </details> <details><summary><b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/cryptiles/package.json,NodeGoat/node_modules/npm/node_modules/request/node_modules/hawk/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/NodeGoat/commit/c221163763b7f2f5d9c526f553b11a21602caa30">c221163763b7f2f5d9c526f553b11a21602caa30</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p> <p>Release Date: 2018-07-09</p> <p>Fix Resolution: v4.1.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"0.2.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;hawk:1.0.0;cryptiles:0.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"},{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"2.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;hawk:3.1.3;cryptiles:2.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1000620","vulnerabilityDetails":"Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1000620 (High) detected in cryptiles-0.2.2.tgz, cryptiles-2.0.5.tgz - ## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>cryptiles-0.2.2.tgz</b>, <b>cryptiles-2.0.5.tgz</b></p></summary> <p> <details><summary><b>cryptiles-0.2.2.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-0.2.2.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-0.2.2.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/zaproxy/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - hawk-1.0.0.tgz - :x: **cryptiles-0.2.2.tgz** (Vulnerable Library) </details> <details><summary><b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/cryptiles/package.json,NodeGoat/node_modules/npm/node_modules/request/node_modules/hawk/node_modules/cryptiles/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/NodeGoat/commit/c221163763b7f2f5d9c526f553b11a21602caa30">c221163763b7f2f5d9c526f553b11a21602caa30</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p> <p>Release Date: 2018-07-09</p> <p>Fix Resolution: v4.1.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"0.2.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;hawk:1.0.0;cryptiles:0.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"},{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"2.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;hawk:3.1.3;cryptiles:2.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-1000620","vulnerabilityDetails":"Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in cryptiles tgz cryptiles tgz cve high severity vulnerability vulnerable libraries cryptiles tgz cryptiles tgz cryptiles tgz general purpose crypto utilities library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules zaproxy node modules cryptiles package json dependency hierarchy zaproxy tgz root library request tgz hawk tgz x cryptiles tgz vulnerable library cryptiles tgz general purpose crypto utilities library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules cryptiles package json nodegoat node modules npm node modules request node modules hawk node modules cryptiles package json dependency hierarchy grunt retire tgz root library request tgz hawk tgz x cryptiles tgz vulnerable library found in head commit a href found in base branch master vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy request hawk cryptiles isminimumfixversionavailable true minimumfixversion packagetype javascript node js packagename cryptiles packageversion packagefilepaths istransitivedependency true dependencytree grunt retire request hawk cryptiles isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in vulnerabilityurl
0
3,551
6,587,491,173
IssuesEvent
2017-09-13 21:17:36
cliffparnitzky/ApparelManager
https://api.github.com/repos/cliffparnitzky/ApparelManager
closed
Show `comment` in lists
Comp - article Comp - article variant Comp - order Comp - order item Improvement ⚙ - Processed
Refers to change in #2 Display the `comment` fields in the list views ... e.g. like with the `product link` field a shortened string
1.0
Show `comment` in lists - Refers to change in #2 Display the `comment` fields in the list views ... e.g. like with the `product link` field a shortened string
process
show comment in lists refers to change in display the comment fields in the list views e g like with the product link field a shortened string
1
456,453
13,150,813,764
IssuesEvent
2020-08-09 13:39:24
chrisjsewell/docutils
https://api.github.com/repos/chrisjsewell/docutils
closed
Caps break ``target`` option in ``image`` directive. [SF:bugs:40]
bugs closed-fixed priority-5
author: ivilata created: 2005-05-15 21:06:41 assigned: felixwiemann SF_url: https://sourceforge.net/p/docutils/bugs/40 The following test input:: .. image:: test.png :target: Target\_ A link to the Target\_. .. \_Target: http://www.example.com/ Results in the following error output:: test.txt:: \(ERROR/3\) Unknown target name: "Target". The first link can not find the name \`\`Target\`\`, while it is defined \(in fact, the second reference works OK\). Changing the argument of \`\`:target:\`\` to \`\`test\`\` makes it work right. Maybe a \`\`tolower\(\)\`\` is lacking somewhere in target handling code. This situation also happens in the \`\`figure\`\` directive. The test is made under SVN HEAD, but also applies to 0.3.7. --- commenter: felixwiemann posted: 2005-05-18 18:18:12 title: #40 Caps break ``target`` option in ``image`` directive. - **status**: open --> closed --- commenter: felixwiemann posted: 2005-05-18 18:18:12 title: #40 Caps break ``target`` option in ``image`` directive. Logged In: YES user\_id=1014490 Fixed; thanks for the bug report. You can download a current snapshot from: http://docutils.sf.net/docutils-snapshot.tgz --- commenter: felixwiemann posted: 2005-05-18 18:18:13 title: #40 Caps break ``target`` option in ``image`` directive. - **assigned_to**: nobody --> felixwiemann - **status**: closed --> closed-fixed
1.0
Caps break ``target`` option in ``image`` directive. [SF:bugs:40] - author: ivilata created: 2005-05-15 21:06:41 assigned: felixwiemann SF_url: https://sourceforge.net/p/docutils/bugs/40 The following test input:: .. image:: test.png :target: Target\_ A link to the Target\_. .. \_Target: http://www.example.com/ Results in the following error output:: test.txt:: \(ERROR/3\) Unknown target name: "Target". The first link can not find the name \`\`Target\`\`, while it is defined \(in fact, the second reference works OK\). Changing the argument of \`\`:target:\`\` to \`\`test\`\` makes it work right. Maybe a \`\`tolower\(\)\`\` is lacking somewhere in target handling code. This situation also happens in the \`\`figure\`\` directive. The test is made under SVN HEAD, but also applies to 0.3.7. --- commenter: felixwiemann posted: 2005-05-18 18:18:12 title: #40 Caps break ``target`` option in ``image`` directive. - **status**: open --> closed --- commenter: felixwiemann posted: 2005-05-18 18:18:12 title: #40 Caps break ``target`` option in ``image`` directive. Logged In: YES user\_id=1014490 Fixed; thanks for the bug report. You can download a current snapshot from: http://docutils.sf.net/docutils-snapshot.tgz --- commenter: felixwiemann posted: 2005-05-18 18:18:13 title: #40 Caps break ``target`` option in ``image`` directive. - **assigned_to**: nobody --> felixwiemann - **status**: closed --> closed-fixed
non_process
caps break target option in image directive author ivilata created assigned felixwiemann sf url the following test input image test png target target a link to the target target results in the following error output test txt error unknown target name target the first link can not find the name target while it is defined in fact the second reference works ok changing the argument of target to test makes it work right maybe a tolower is lacking somewhere in target handling code this situation also happens in the figure directive the test is made under svn head but also applies to commenter felixwiemann posted title caps break target option in image directive status open closed commenter felixwiemann posted title caps break target option in image directive logged in yes user id fixed thanks for the bug report you can download a current snapshot from commenter felixwiemann posted title caps break target option in image directive assigned to nobody felixwiemann status closed closed fixed
0
1,461
2,546,130,298
IssuesEvent
2015-01-29 21:42:12
SU-SWS/open_framework
https://api.github.com/repos/SU-SWS/open_framework
closed
Remove floats from how we implement postcards
7.x-2.x CLIENT RISK enhancement NEEDS TESTING solution found!
This is a problem when the text is shorter than the image, and because the image is floating, content below the postcard will bump up. We've solved this with hacks for various views by putting min-heights on the views-row to force them to be at least as tall as the image. This is a bad way to solve this problem. Can we implement postcards without a float in the styles?
1.0
Remove floats from how we implement postcards - This is a problem when the text is shorter than the image, and because the image is floating, content below the postcard will bump up. We've solved this with hacks for various views by putting min-heights on the views-row to force them to be at least as tall as the image. This is a bad way to solve this problem. Can we implement postcards without a float in the styles?
non_process
remove floats from how we implement postcards this is a problem when the text is shorter than the image and because the image is floating content below the postcard will bump up we ve solved this with hacks for various views by putting min heights on the views row to force them to be at least as tall as the image this is a bad way to solve this problem can we implement postcards without a float in the styles
0
20,133
26,676,436,616
IssuesEvent
2023-01-26 14:36:08
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
GUI to define the BPMN app process
kind/feature-request area/process area/ui-editor Epic status/duplicate ux
## Is your feature request related to a problem? Please describe Defining the process for the app can be un-intuitive and the threshold to do so is higher than necessary. ## Describe the solution you'd like A GUI to define the process for the app ## Describe alternatives you've considered A library of pre-defined processes. ## Additional context Financing of this issue is granted through Medfinansieringsordningen.
1.0
GUI to define the BPMN app process - ## Is your feature request related to a problem? Please describe Defining the process for the app can be un-intuitive and the threshold to do so is higher than necessary. ## Describe the solution you'd like A GUI to define the process for the app ## Describe alternatives you've considered A library of pre-defined processes. ## Additional context Financing of this issue is granted through Medfinansieringsordningen.
process
gui to define the bpmn app process is your feature request related to a problem please describe defining the process for the app can be un intuitive and the threshold to do so is higher than necessary describe the solution you d like a gui to define the process for the app describe alternatives you ve considered a library of pre defined processes additional context financing of this issue is granted through medfinansieringsordningen
1
78,390
15,569,982,219
IssuesEvent
2021-03-17 01:27:00
jrrk/riscv-linux
https://api.github.com/repos/jrrk/riscv-linux
opened
CVE-2020-12652 (Medium) detected in aspeedaspeed-4.19-devicetree-no-fsi
security vulnerability
## CVE-2020-12652 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>aspeedaspeed-4.19-devicetree-no-fsi</b></p></summary> <p> <p>ASPEED ARM SoC development</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/joel/aspeed.git>https://git.kernel.org/pub/scm/linux/kernel/git/joel/aspeed.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>riscv-linux/drivers/message/fusion/mptctl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>riscv-linux/drivers/message/fusion/mptctl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The __mptctl_ioctl function in drivers/message/fusion/mptctl.c in the Linux kernel before 5.4.14 allows local users to hold an incorrect lock during the ioctl operation and trigger a race condition, i.e., a "double fetch" vulnerability, aka CID-28d76df18f0a. NOTE: the vendor states "The security impact of this bug is not as bad as it could have been because these operations are all privileged and root already has enormous destructive power." <p>Publish Date: 2020-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12652>CVE-2020-12652</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.4.14">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.4.14</a></p> <p>Release Date: 2020-05-05</p> <p>Fix Resolution: 5.4.14</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-12652 (Medium) detected in aspeedaspeed-4.19-devicetree-no-fsi - ## CVE-2020-12652 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>aspeedaspeed-4.19-devicetree-no-fsi</b></p></summary> <p> <p>ASPEED ARM SoC development</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/joel/aspeed.git>https://git.kernel.org/pub/scm/linux/kernel/git/joel/aspeed.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>riscv-linux/drivers/message/fusion/mptctl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>riscv-linux/drivers/message/fusion/mptctl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The __mptctl_ioctl function in drivers/message/fusion/mptctl.c in the Linux kernel before 5.4.14 allows local users to hold an incorrect lock during the ioctl operation and trigger a race condition, i.e., a "double fetch" vulnerability, aka CID-28d76df18f0a. NOTE: the vendor states "The security impact of this bug is not as bad as it could have been because these operations are all privileged and root already has enormous destructive power." <p>Publish Date: 2020-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12652>CVE-2020-12652</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.4.14">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.4.14</a></p> <p>Release Date: 2020-05-05</p> <p>Fix Resolution: 5.4.14</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in aspeedaspeed devicetree no fsi cve medium severity vulnerability vulnerable library aspeedaspeed devicetree no fsi aspeed arm soc development library home page a href vulnerable source files riscv linux drivers message fusion mptctl c riscv linux drivers message fusion mptctl c vulnerability details the mptctl ioctl function in drivers message fusion mptctl c in the linux kernel before allows local users to hold an incorrect lock during the ioctl operation and trigger a race condition i e a double fetch vulnerability aka cid note the vendor states the security impact of this bug is not as bad as it could have been because these operations are all privileged and root already has enormous destructive power publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
5,558
8,395,049,003
IssuesEvent
2018-10-10 04:12:03
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Parse results are cached even when references are changed
bug parse-tree-processing performance
Adding, removing, or changing the order of a reference can alter the resolver results for a module. The parser seems to skip parsing a module if it is unchanged, even though a reference change might make a substantial impact on resolution. The parser should take reference changes into account when caching results.
1.0
Parse results are cached even when references are changed - Adding, removing, or changing the order of a reference can alter the resolver results for a module. The parser seems to skip parsing a module if it is unchanged, even though a reference change might make a substantial impact on resolution. The parser should take reference changes into account when caching results.
process
parse results are cached even when references are changed adding removing or changing the order of a reference can alter the resolver results for a module the parser seems to skip parsing a module if it is unchanged even though a reference change might make a substantial impact on resolution the parser should take reference changes into account when caching results
1
9,091
3,254,362,227
IssuesEvent
2015-10-19 23:37:01
twosigma/beaker-notebook
https://api.github.com/repos/twosigma/beaker-notebook
closed
improve error reporting when wrong python runs
Bug Core Client Documentation Language Plugins
Problem: when open a default worksheet with ipython and R, there is error popup "error no module named 'urllib2' " BTW: looks like ipython is still working by typing "1+1", "1*6+7" Environment: OS: Win 7 64 bit browser: google chorme notebook version: 1.4.1-0-ge8427bf-windows Log information D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Copyrigh t 2014 TWO SIGMA OPEN SOURCE, LLC D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Licensed under the Apache License, Version 2.0 (the "License"); D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM you may not use this file except in compliance with the License. D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM You may obtain a copy of the License at D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM h ttp://www.apache.org/licenses/LICENSE-2.0 D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Unless r equired by applicable law or agreed to in writing, software D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM distribu ted under the License is distributed on an "AS IS" BASIS, D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM See the License for the specific language governing permissions and D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM limitati ons under the License. Port range 8800-8803 taken, searching... Oct 14, 2015 1:58:11 PM com.twosigma.beaker.shared.module.util.GeneralUtilsImpl readFile INFO: ERROR reading file C:\Users\H141930\.beaker\v1\config\beaker.pref.json starting nginx instance (D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge84 27bf-windows/nginx) [main] INFO org.eclipse.jetty.server.Server - jetty-8.1.13.v20130916 [main] INFO org.eclipse.jetty.server.AbstractConnector - Started SelectChannelCo nnector@127.0.0.1:8806 Connecting to http://127.0.0.1:8805/ Running 0: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows\python\py thon 1: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows/config/pl ugins/eval/ipythonPlugins/ipython/ipythonPlugin 2: 8808 Done starting IPython Running 0: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows\python\py thon 1: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows/config/pl ugins/eval/r/rPlugin 2: 8809 Done starting R
1.0
improve error reporting when wrong python runs - Problem: when open a default worksheet with ipython and R, there is error popup "error no module named 'urllib2' " BTW: looks like ipython is still working by typing "1+1", "1*6+7" Environment: OS: Win 7 64 bit browser: google chorme notebook version: 1.4.1-0-ge8427bf-windows Log information D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Copyrigh t 2014 TWO SIGMA OPEN SOURCE, LLC D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Licensed under the Apache License, Version 2.0 (the "License"); D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM you may not use this file except in compliance with the License. D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM You may obtain a copy of the License at D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM h ttp://www.apache.org/licenses/LICENSE-2.0 D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM Unless r equired by applicable law or agreed to in writing, software D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM distribu ted under the License is distributed on an "AS IS" BASIS, D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM See the License for the specific language governing permissions and D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows>REM limitati ons under the License. Port range 8800-8803 taken, searching... Oct 14, 2015 1:58:11 PM com.twosigma.beaker.shared.module.util.GeneralUtilsImpl readFile INFO: ERROR reading file C:\Users\H141930\.beaker\v1\config\beaker.pref.json starting nginx instance (D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge84 27bf-windows/nginx) [main] INFO org.eclipse.jetty.server.Server - jetty-8.1.13.v20130916 [main] INFO org.eclipse.jetty.server.AbstractConnector - Started SelectChannelCo nnector@127.0.0.1:8806 Connecting to http://127.0.0.1:8805/ Running 0: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows\python\py thon 1: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows/config/pl ugins/eval/ipythonPlugins/ipython/ipythonPlugin 2: 8808 Done starting IPython Running 0: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows\python\py thon 1: D:\Software\BeakerNotebook\beaker-notebook-1.4.1-0-ge8427bf-windows/config/pl ugins/eval/r/rPlugin 2: 8809 Done starting R
non_process
improve error reporting when wrong python runs problem when open a default worksheet with ipython and r there is error popup error no module named btw looks like ipython is still working by typing environment os win bit browser google chorme notebook version windows log information d software beakernotebook beaker notebook windows rem d software beakernotebook beaker notebook windows rem copyrigh t two sigma open source llc d software beakernotebook beaker notebook windows rem d software beakernotebook beaker notebook windows rem licensed under the apache license version the license d software beakernotebook beaker notebook windows rem you may not use this file except in compliance with the license d software beakernotebook beaker notebook windows rem you may obtain a copy of the license at d software beakernotebook beaker notebook windows rem d software beakernotebook beaker notebook windows rem h ttp d software beakernotebook beaker notebook windows rem d software beakernotebook beaker notebook windows rem unless r equired by applicable law or agreed to in writing software d software beakernotebook beaker notebook windows rem distribu ted under the license is distributed on an as is basis d software beakernotebook beaker notebook windows rem without warranties or conditions of any kind either express or implied d software beakernotebook beaker notebook windows rem see the license for the specific language governing permissions and d software beakernotebook beaker notebook windows rem limitati ons under the license port range taken searching oct pm com twosigma beaker shared module util generalutilsimpl readfile info error reading file c users beaker config beaker pref json starting nginx instance d software beakernotebook beaker notebook windows nginx info org eclipse jetty server server jetty info org eclipse jetty server abstractconnector started selectchannelco nnector connecting to running d software beakernotebook beaker notebook windows python py thon d software beakernotebook beaker notebook windows config pl ugins eval ipythonplugins ipython ipythonplugin done starting ipython running d software beakernotebook beaker notebook windows python py thon d software beakernotebook beaker notebook windows config pl ugins eval r rplugin done starting r
0
54,992
13,943,655,645
IssuesEvent
2020-10-22 23:42:44
mlmcd/WebGoat
https://api.github.com/repos/mlmcd/WebGoat
opened
CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sass0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70
security vulnerability
## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sass0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70</b></p></summary> <p> <details><summary><b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - gulp-sass-4.0.2.tgz (Root Library) - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: Libsass-3.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.11.0","isTransitiveDependency":true,"dependencyTree":"gulp-sass:4.0.2;node-sass:4.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Libsass-3.6.0"}],"vulnerabilityIdentifier":"CVE-2018-11698","vulnerabilityDetails":"An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sass0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70 - ## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sass0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70</b></p></summary> <p> <details><summary><b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - gulp-sass-4.0.2.tgz (Root Library) - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-11698</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: Libsass-3.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.11.0","isTransitiveDependency":true,"dependencyTree":"gulp-sass:4.0.2;node-sass:4.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Libsass-3.6.0"}],"vulnerabilityIdentifier":"CVE-2018-11698","vulnerabilityDetails":"An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in node sass tgz node cve high severity vulnerability vulnerable libraries node sass tgz node node sass tgz wrapper around libsass library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules node sass package json dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch develop vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service vulnerabilityurl
0
105,915
13,231,692,009
IssuesEvent
2020-08-18 12:08:21
carbon-design-system/digital-design-ideation
https://api.github.com/repos/carbon-design-system/digital-design-ideation
opened
Lead space: Variations Final design
Design Final design
### User Story > As an [user role below]: `ibm.com user` > I would like to be able to: `orient myself and understand the content of the page` > so that I can: `make a decision on whether this page is relevant to me or not.` ### user story 2 > As an [user role below]: `DDS designer` > I would like to be able to: `understand what kind of lead spaces are out there both internally and externally` > so that I can: `create a system of lead spaces.` #### Additional Information - **See the epic for the overarching descriptions of the work.** - [box folder] - [reference materials] #### Acceptance Criteria - [ ] Final designs are available - [ ] Design spec issue has been created on the Agile board with the documentation needed and the work scheduled.
2.0
Lead space: Variations Final design - ### User Story > As an [user role below]: `ibm.com user` > I would like to be able to: `orient myself and understand the content of the page` > so that I can: `make a decision on whether this page is relevant to me or not.` ### user story 2 > As an [user role below]: `DDS designer` > I would like to be able to: `understand what kind of lead spaces are out there both internally and externally` > so that I can: `create a system of lead spaces.` #### Additional Information - **See the epic for the overarching descriptions of the work.** - [box folder] - [reference materials] #### Acceptance Criteria - [ ] Final designs are available - [ ] Design spec issue has been created on the Agile board with the documentation needed and the work scheduled.
non_process
lead space variations final design user story as an ibm com user i would like to be able to orient myself and understand the content of the page so that i can make a decision on whether this page is relevant to me or not user story as an dds designer i would like to be able to understand what kind of lead spaces are out there both internally and externally so that i can create a system of lead spaces additional information see the epic for the overarching descriptions of the work acceptance criteria final designs are available design spec issue has been created on the agile board with the documentation needed and the work scheduled
0
144,748
13,124,390,611
IssuesEvent
2020-08-06 03:36:33
up-banking/api
https://api.github.com/repos/up-banking/api
closed
Webhook Ping returns an error
bug documentation
When calling Ping url via Curl (as per example) `curl https://api.up.com.au/api/v1/webhooks/316c2282-4b75-4539-9561-1893564d4dad/ping \ > -XPOST \ > -H 'Authorization: Bearer up:yeah ...` it returns with an error like: `<!DOCTYPE html> <html lang=en> <meta charset=utf-8> <meta name=viewport content="initial-scale=1, minimum-scale=1, width=device-width"> <title>Error 411 (Length Required)!!1</title> <style> *{margin:0;padding:0}html,code{font:15px/22px arial,sans-serif}html{background:#fff;color:#222;padding:15px}body{margin:7% auto 0;max-width:390px;min-height:180px;padding:30px 0 15px}* > body{background:url(//www.google.com/images/errors/robot.png) 100% 5px no-repeat;padding-right:205px}p{margin:11px 0 22px;overflow:hidden}ins{color:#777;text-decoration:none}a img{border:0}@media screen and (max-width:772px){body{background:none;margin-top:0;max-width:none;padding-right:0}}#logo{background:url(//www.google.com/images/branding/googlelogo/1x/googlelogo_color_150x54dp.png) no-repeat;margin-left:-5px}@media only screen and (min-resolution:192dpi){#logo{background:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) no-repeat 0% 0%/100% 100%;-moz-border-image:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) 0}}@media only screen and (-webkit-min-device-pixel-ratio:2){#logo{background:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) no-repeat;-webkit-background-size:100% 100%}}#logo{display:inline-block;height:54px;width:150px} </style> <a href=//www.google.com/><span id=logo aria-label=Google></span></a> <p><b>411.</b> <ins>That’s an error.</ins> <p>POST requests require a <code>Content-length</code> header. <ins>That’s all we know.</ins>`
1.0
Webhook Ping returns an error - When calling Ping url via Curl (as per example) `curl https://api.up.com.au/api/v1/webhooks/316c2282-4b75-4539-9561-1893564d4dad/ping \ > -XPOST \ > -H 'Authorization: Bearer up:yeah ...` it returns with an error like: `<!DOCTYPE html> <html lang=en> <meta charset=utf-8> <meta name=viewport content="initial-scale=1, minimum-scale=1, width=device-width"> <title>Error 411 (Length Required)!!1</title> <style> *{margin:0;padding:0}html,code{font:15px/22px arial,sans-serif}html{background:#fff;color:#222;padding:15px}body{margin:7% auto 0;max-width:390px;min-height:180px;padding:30px 0 15px}* > body{background:url(//www.google.com/images/errors/robot.png) 100% 5px no-repeat;padding-right:205px}p{margin:11px 0 22px;overflow:hidden}ins{color:#777;text-decoration:none}a img{border:0}@media screen and (max-width:772px){body{background:none;margin-top:0;max-width:none;padding-right:0}}#logo{background:url(//www.google.com/images/branding/googlelogo/1x/googlelogo_color_150x54dp.png) no-repeat;margin-left:-5px}@media only screen and (min-resolution:192dpi){#logo{background:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) no-repeat 0% 0%/100% 100%;-moz-border-image:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) 0}}@media only screen and (-webkit-min-device-pixel-ratio:2){#logo{background:url(//www.google.com/images/branding/googlelogo/2x/googlelogo_color_150x54dp.png) no-repeat;-webkit-background-size:100% 100%}}#logo{display:inline-block;height:54px;width:150px} </style> <a href=//www.google.com/><span id=logo aria-label=Google></span></a> <p><b>411.</b> <ins>That’s an error.</ins> <p>POST requests require a <code>Content-length</code> header. <ins>That’s all we know.</ins>`
non_process
webhook ping returns an error when calling ping url via curl as per example curl xpost h authorization bearer up yeah it returns with an error like error length required margin padding html code font arial sans serif html background fff color padding body margin auto max width min height padding body background url no repeat padding right p margin overflow hidden ins color text decoration none a img border media screen and max width body background none margin top max width none padding right logo background url no repeat margin left media only screen and min resolution logo background url no repeat moz border image url media only screen and webkit min device pixel ratio logo background url no repeat webkit background size logo display inline block height width that’s an error post requests require a content length header that’s all we know
0
1,813
4,561,550,179
IssuesEvent
2016-09-14 12:10:42
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
grouping shocks in decompositions
enhancement preprocessor
would it be possible to allow a more flexible naming of the groups, allowing a sort of verbatim reading of what is before `=` . the following syntax does not work ``` shock_groups; ... Price Mark-up EA = emup; Wage Mark-up EA = emuw; ... end; ``` not sure if this can be modified, by allowing group names to be given inside brackets, as follows: ``` shock_groups; ... (Price Mark-up EA) = emup; (Wage Mark-up EA) = emuw; ... end; ``` ?
1.0
grouping shocks in decompositions - would it be possible to allow a more flexible naming of the groups, allowing a sort of verbatim reading of what is before `=` . the following syntax does not work ``` shock_groups; ... Price Mark-up EA = emup; Wage Mark-up EA = emuw; ... end; ``` not sure if this can be modified, by allowing group names to be given inside brackets, as follows: ``` shock_groups; ... (Price Mark-up EA) = emup; (Wage Mark-up EA) = emuw; ... end; ``` ?
process
grouping shocks in decompositions would it be possible to allow a more flexible naming of the groups allowing a sort of verbatim reading of what is before the following syntax does not work shock groups price mark up ea emup wage mark up ea emuw end not sure if this can be modified by allowing group names to be given inside brackets as follows shock groups price mark up ea emup wage mark up ea emuw end
1
15,957
20,173,708,200
IssuesEvent
2022-02-10 12:45:26
ooi-data/CE06ISSM-MFD37-01-OPTAAD000-telemetered-optaa_dj_dcl_instrument
https://api.github.com/repos/ooi-data/CE06ISSM-MFD37-01-OPTAAD000-telemetered-optaa_dj_dcl_instrument
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-02-10T12:45:25.184357. ## Details Flow name: `CE06ISSM-MFD37-01-OPTAAD000-telemetered-optaa_dj_dcl_instrument` Task name: `processing_task` Error type: `ValueError` Error message: cannot reindex or align along dimension 'wavelength' because the index has duplicate values <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 355, in append_to_zarr mod_ds = mod_ds.reindex(dim_indexer) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2947, in reindex return self._reindex( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2976, in _reindex variables, indexes = alignment.reindex_variables( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/alignment.py", line 589, in reindex_variables raise ValueError( ValueError: cannot reindex or align along dimension 'wavelength' because the index has duplicate values ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-02-10T12:45:25.184357. ## Details Flow name: `CE06ISSM-MFD37-01-OPTAAD000-telemetered-optaa_dj_dcl_instrument` Task name: `processing_task` Error type: `ValueError` Error message: cannot reindex or align along dimension 'wavelength' because the index has duplicate values <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 355, in append_to_zarr mod_ds = mod_ds.reindex(dim_indexer) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2947, in reindex return self._reindex( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2976, in _reindex variables, indexes = alignment.reindex_variables( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/alignment.py", line 589, in reindex_variables raise ValueError( ValueError: cannot reindex or align along dimension 'wavelength' because the index has duplicate values ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name telemetered optaa dj dcl instrument task name processing task error type valueerror error message cannot reindex or align along dimension wavelength because the index has duplicate values traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing process dataset file srv conda envs notebook lib site packages ooi harvester processor init py line in process dataset append to zarr mod ds store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr mod ds mod ds reindex dim indexer file srv conda envs notebook lib site packages xarray core dataset py line in reindex return self reindex file srv conda envs notebook lib site packages xarray core dataset py line in reindex variables indexes alignment reindex variables file srv conda envs notebook lib site packages xarray core alignment py line in reindex variables raise valueerror valueerror cannot reindex or align along dimension wavelength because the index has duplicate values
1
783,474
27,532,650,731
IssuesEvent
2023-03-06 23:38:57
HydrologicEngineeringCenter/HEC-FDA
https://api.github.com/repos/HydrologicEngineeringCenter/HEC-FDA
reopened
subtract quantiles
bug PRIORITY
We need to re-wire the way that we combine distributions so that we are combining quantiles instead of random independent samples.
1.0
subtract quantiles - We need to re-wire the way that we combine distributions so that we are combining quantiles instead of random independent samples.
non_process
subtract quantiles we need to re wire the way that we combine distributions so that we are combining quantiles instead of random independent samples
0
12,165
14,741,557,370
IssuesEvent
2021-01-07 10:48:23
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
opened
Create a universal log Project
anc-core anp-1.5 ant-feature grt-ui processes has attachment pl-foran
In GitLab by @kdjstudios on Jan 24, 2019, 09:35 We need to research into the what it would take to create a General User Audit Log. - This would be happening through the UI and be triggered via JavaScript hooks. All tickets currently open that request this type of functionality: * #886 * #87 * #1427 * #166 * #732 * #776 # Logging: Below will be the information that will be saved into a database of log entries. These log entries will then be pulled by a report/UI to display them to the users. Each "button" in SAB will basically have this "logging" function that it uses, which will save everything that occurred. This means that for editing a site, when the "Save" button is clicked all edits made will need their own 'line item'; such if I edited three activities, each activity edit will have it's own line in the log. This should be done via JavaScript hooks as described above, unless there is a better way to implement this. ### Information logged for each 'System Event'. * Date and Time in UTC: This is the time the user performed the action. * Action: This is the action taken by the system. * Old State: What the value was before the user adjusted it. * New State: What the value is now after the user adjusted it. ### Information logged for each 'User Event'. * Date and Time in UTC: This is the time the user performed the action. * Username: This is the username of the action performed. * Action: This is the action taken by the user. Generally the name of the button clicked. * Action Type: This is the section or location of the action. Generally will provide where the user was in the system. * Site: This will be site name associated with the action. * Account: This will be account name associated with the action. * Customer: This will be customer name associated with the action. * Field Name: This is the field name/label. * Old State: What the value was before the user adjusted it. * New State: What the value is now after the user adjusted it. # Mock Up Report: Here is a mock up of the report and of all the various items we need logged. Basically every button click that is available in SAB we will need to log. The description field, will have more details and this should be a field that is generated and saved directly into the DB at the time of the logging of the action. This way each button can "format" this as needed. This will also insure there is limited adjustments to the data when running the report. The report should run by default for the last 7 days. ![Log_Report](/uploads/9496f0a7f085f067bf6526312145acfa/Log_Report.png)
1.0
Create a universal log Project - In GitLab by @kdjstudios on Jan 24, 2019, 09:35 We need to research into the what it would take to create a General User Audit Log. - This would be happening through the UI and be triggered via JavaScript hooks. All tickets currently open that request this type of functionality: * #886 * #87 * #1427 * #166 * #732 * #776 # Logging: Below will be the information that will be saved into a database of log entries. These log entries will then be pulled by a report/UI to display them to the users. Each "button" in SAB will basically have this "logging" function that it uses, which will save everything that occurred. This means that for editing a site, when the "Save" button is clicked all edits made will need their own 'line item'; such if I edited three activities, each activity edit will have it's own line in the log. This should be done via JavaScript hooks as described above, unless there is a better way to implement this. ### Information logged for each 'System Event'. * Date and Time in UTC: This is the time the user performed the action. * Action: This is the action taken by the system. * Old State: What the value was before the user adjusted it. * New State: What the value is now after the user adjusted it. ### Information logged for each 'User Event'. * Date and Time in UTC: This is the time the user performed the action. * Username: This is the username of the action performed. * Action: This is the action taken by the user. Generally the name of the button clicked. * Action Type: This is the section or location of the action. Generally will provide where the user was in the system. * Site: This will be site name associated with the action. * Account: This will be account name associated with the action. * Customer: This will be customer name associated with the action. * Field Name: This is the field name/label. * Old State: What the value was before the user adjusted it. * New State: What the value is now after the user adjusted it. # Mock Up Report: Here is a mock up of the report and of all the various items we need logged. Basically every button click that is available in SAB we will need to log. The description field, will have more details and this should be a field that is generated and saved directly into the DB at the time of the logging of the action. This way each button can "format" this as needed. This will also insure there is limited adjustments to the data when running the report. The report should run by default for the last 7 days. ![Log_Report](/uploads/9496f0a7f085f067bf6526312145acfa/Log_Report.png)
process
create a universal log project in gitlab by kdjstudios on jan we need to research into the what it would take to create a general user audit log this would be happening through the ui and be triggered via javascript hooks all tickets currently open that request this type of functionality logging below will be the information that will be saved into a database of log entries these log entries will then be pulled by a report ui to display them to the users each button in sab will basically have this logging function that it uses which will save everything that occurred this means that for editing a site when the save button is clicked all edits made will need their own line item such if i edited three activities each activity edit will have it s own line in the log this should be done via javascript hooks as described above unless there is a better way to implement this information logged for each system event date and time in utc this is the time the user performed the action action this is the action taken by the system old state what the value was before the user adjusted it new state what the value is now after the user adjusted it information logged for each user event date and time in utc this is the time the user performed the action username this is the username of the action performed action this is the action taken by the user generally the name of the button clicked action type this is the section or location of the action generally will provide where the user was in the system site this will be site name associated with the action account this will be account name associated with the action customer this will be customer name associated with the action field name this is the field name label old state what the value was before the user adjusted it new state what the value is now after the user adjusted it mock up report here is a mock up of the report and of all the various items we need logged basically every button click that is available in sab we will need to log the description field will have more details and this should be a field that is generated and saved directly into the db at the time of the logging of the action this way each button can format this as needed this will also insure there is limited adjustments to the data when running the report the report should run by default for the last days uploads log report png
1