Unnamed: 0
int64
9
832k
id
float64
2.5B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
4
323
labels
stringlengths
4
2.67k
body
stringlengths
23
107k
index
stringclasses
4 values
text_combine
stringlengths
96
107k
label
stringclasses
2 values
text
stringlengths
96
56.1k
binary_label
int64
0
1
1,415
15,964,711,126
IssuesEvent
2021-04-16 06:42:13
emmamei/cdkey
https://api.github.com/repos/emmamei/cdkey
opened
setDollType function doesn't check for valid type
bug reliabilityfix
When `setDollType` is called, it doesn't check for a valid Type. This needs to be fixed.
True
setDollType function doesn't check for valid type - When `setDollType` is called, it doesn't check for a valid Type. This needs to be fixed.
reli
setdolltype function doesn t check for valid type when setdolltype is called it doesn t check for a valid type this needs to be fixed
1
2,267
24,562,137,570
IssuesEvent
2022-10-12 21:24:02
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
opened
[FEA] Should we synchronize and then spill with the ASYNC allocator
feature request ? - Needs Triage performance reliability
As mentioned here https://github.com/NVIDIA/spark-rapids/issues/6768, I am noticing the synchronizing on OOM can help us handle allocation failures that would otherwise be fatal. Additionally, with some quick prototyping locally, it seems that there may be a performance gain here. Specifically, if we first `Cuda.deviceSynchronize` rather than spill right away, but fallback to the spill when we know we have already synchronized, we are able to save time with a quick query I tried in our performance cluster. I ran a query that spills constantly and it took 265 seconds vs 304 seconds without this change. That said the query also ran OOM on a second trial. The reason I think is that we are really able to pack the GPU, I see that the async pool is able to get closer to its maximum size (40GB in this case). So we have less fudge memory for those tasks that run above their ~1/concurrentGpuTasks chunk of memory.
True
[FEA] Should we synchronize and then spill with the ASYNC allocator - As mentioned here https://github.com/NVIDIA/spark-rapids/issues/6768, I am noticing the synchronizing on OOM can help us handle allocation failures that would otherwise be fatal. Additionally, with some quick prototyping locally, it seems that there may be a performance gain here. Specifically, if we first `Cuda.deviceSynchronize` rather than spill right away, but fallback to the spill when we know we have already synchronized, we are able to save time with a quick query I tried in our performance cluster. I ran a query that spills constantly and it took 265 seconds vs 304 seconds without this change. That said the query also ran OOM on a second trial. The reason I think is that we are really able to pack the GPU, I see that the async pool is able to get closer to its maximum size (40GB in this case). So we have less fudge memory for those tasks that run above their ~1/concurrentGpuTasks chunk of memory.
reli
should we synchronize and then spill with the async allocator as mentioned here i am noticing the synchronizing on oom can help us handle allocation failures that would otherwise be fatal additionally with some quick prototyping locally it seems that there may be a performance gain here specifically if we first cuda devicesynchronize rather than spill right away but fallback to the spill when we know we have already synchronized we are able to save time with a quick query i tried in our performance cluster i ran a query that spills constantly and it took seconds vs seconds without this change that said the query also ran oom on a second trial the reason i think is that we are really able to pack the gpu i see that the async pool is able to get closer to its maximum size in this case so we have less fudge memory for those tasks that run above their concurrentgputasks chunk of memory
1
1,645
18,050,901,285
IssuesEvent
2021-09-19 18:25:55
ruuseljo/schoolprojectCO2
https://api.github.com/repos/ruuseljo/schoolprojectCO2
opened
the program should auto re-start if it has died
reliability
use a cron job to periodically check the processes for the main process. if it doesn't exist, re-start the process. if it does exist and is a zombie, kill and re-start. if it does exist and is not zombie, we are as happy as Larry.
True
the program should auto re-start if it has died - use a cron job to periodically check the processes for the main process. if it doesn't exist, re-start the process. if it does exist and is a zombie, kill and re-start. if it does exist and is not zombie, we are as happy as Larry.
reli
the program should auto re start if it has died use a cron job to periodically check the processes for the main process if it doesn t exist re start the process if it does exist and is a zombie kill and re start if it does exist and is not zombie we are as happy as larry
1
454,154
13,095,710,533
IssuesEvent
2020-08-03 14:32:31
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Unit tests are broken
priority: medium quality
## Describe the bug Studio unit tests are broken in develop & support/3.1.x branchs ## To Reproduce Steps to reproduce the behavior: 1. Clone studio develop branch or `./gradlew init` 2. Go into `src/studio` and run `mvn clean test` 3. See errors ## Expected behavior All tests pass without error. ## Screenshots N/A ## Logs ``` studio] mvn test [INFO] Scanning for projects... [INFO] [INFO] -------------------< org.craftercms:crafter-studio >-------------------- [INFO] Building Crafter Studio 3.2.0-SNAPSHOT [INFO] --------------------------------[ war ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ crafter-studio --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 218 resources [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create (Java Code) @ crafter-studio --- [INFO] Checking for local modifications: skipped. [INFO] Updating project files from SCM: skipped. [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio [INFO] Storing buildNumber: a9ff303f84626f5ff74e51536c4be166a1aacd05 at timestamp: 1592499185453 [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio [INFO] Storing buildScmBranch: UNKNOWN [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (Java Code) @ crafter-studio --- [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create (UI-Version) @ crafter-studio --- [INFO] Checking for local modifications: skipped. [INFO] Updating project files from SCM: skipped. [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui [WARNING] Cannot get the revision information from the scm repository, proceeding with revision of unknown : Exception while executing SCM command. [INFO] Storing buildNumber: unknown at timestamp: 1592499185630 [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui [WARNING] Cannot get the branch information from the scm repository, proceeding with UNKNOWN_BRANCH : Exception while executing SCM command. [INFO] Storing buildScmBranch: UNKNOWN_BRANCH [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (UI-Version) @ crafter-studio --- [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ crafter-studio --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ crafter-studio --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 4 resources [INFO] [INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ crafter-studio --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 5 source files to /home/jose/github/craftercms/src/studio/target/test-classes [INFO] ------------------------------------------------------------- [ERROR] COMPILATION ERROR : [INFO] ------------------------------------------------------------- [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int) location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository [INFO] 5 errors [INFO] ------------------------------------------------------------- [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 2.061 s [INFO] Finished at: 2020-06-18T10:53:06-06:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project crafter-studio: Compilation failure: Compilation failure: [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol [ERROR] symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int) [ERROR] location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException ``` ## Specs ### Version Studio Develop branch Studio Support/3.1.x branch ### OS Linux ### Browser N/A ## Additional context N/A
1.0
[studio] Unit tests are broken - ## Describe the bug Studio unit tests are broken in develop & support/3.1.x branchs ## To Reproduce Steps to reproduce the behavior: 1. Clone studio develop branch or `./gradlew init` 2. Go into `src/studio` and run `mvn clean test` 3. See errors ## Expected behavior All tests pass without error. ## Screenshots N/A ## Logs ``` studio] mvn test [INFO] Scanning for projects... [INFO] [INFO] -------------------< org.craftercms:crafter-studio >-------------------- [INFO] Building Crafter Studio 3.2.0-SNAPSHOT [INFO] --------------------------------[ war ]--------------------------------- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ crafter-studio --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 218 resources [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create (Java Code) @ crafter-studio --- [INFO] Checking for local modifications: skipped. [INFO] Updating project files from SCM: skipped. [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio [INFO] Storing buildNumber: a9ff303f84626f5ff74e51536c4be166a1aacd05 at timestamp: 1592499185453 [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio [INFO] Storing buildScmBranch: UNKNOWN [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (Java Code) @ crafter-studio --- [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create (UI-Version) @ crafter-studio --- [INFO] Checking for local modifications: skipped. [INFO] Updating project files from SCM: skipped. [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui [WARNING] Cannot get the revision information from the scm repository, proceeding with revision of unknown : Exception while executing SCM command. [INFO] Storing buildNumber: unknown at timestamp: 1592499185630 [INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD [INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui [WARNING] Cannot get the branch information from the scm repository, proceeding with UNKNOWN_BRANCH : Exception while executing SCM command. [INFO] Storing buildScmBranch: UNKNOWN_BRANCH [INFO] [INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (UI-Version) @ crafter-studio --- [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ crafter-studio --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ crafter-studio --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 4 resources [INFO] [INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ crafter-studio --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 5 source files to /home/jose/github/craftercms/src/studio/target/test-classes [INFO] ------------------------------------------------------------- [ERROR] COMPILATION ERROR : [INFO] ------------------------------------------------------------- [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int) location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository [INFO] 5 errors [INFO] ------------------------------------------------------------- [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 2.061 s [INFO] Finished at: 2020-06-18T10:53:06-06:00 [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project crafter-studio: Compilation failure: Compilation failure: [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol [ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String) [ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol [ERROR] symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int) [ERROR] location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException ``` ## Specs ### Version Studio Develop branch Studio Support/3.1.x branch ### OS Linux ### Browser N/A ## Additional context N/A
non_reli
unit tests are broken describe the bug studio unit tests are broken in develop support x branchs to reproduce steps to reproduce the behavior clone studio develop branch or gradlew init go into src studio and run mvn clean test see errors expected behavior all tests pass without error screenshots n a logs studio mvn test scanning for projects building crafter studio snapshot maven resources plugin resources default resources crafter studio using utf encoding to copy filtered resources copying resources buildnumber maven plugin create java code crafter studio checking for local modifications skipped updating project files from scm skipped executing bin sh c cd home jose github craftercms src studio git rev parse verify head working directory home jose github craftercms src studio storing buildnumber at timestamp executing bin sh c cd home jose github craftercms src studio git rev parse verify head working directory home jose github craftercms src studio storing buildscmbranch unknown buildnumber maven plugin create timestamp java code crafter studio buildnumber maven plugin create ui version crafter studio checking for local modifications skipped updating project files from scm skipped executing bin sh c cd home jose github craftercms src studio target ui git rev parse verify head working directory home jose github craftercms src studio target ui cannot get the revision information from the scm repository proceeding with revision of unknown exception while executing scm command storing buildnumber unknown at timestamp executing bin sh c cd home jose github craftercms src studio target ui git rev parse verify head working directory home jose github craftercms src studio target ui cannot get the branch information from the scm repository proceeding with unknown branch exception while executing scm command storing buildscmbranch unknown branch buildnumber maven plugin create timestamp ui version crafter studio maven compiler plugin compile default compile crafter studio nothing to compile all classes are up to date maven resources plugin testresources default testresources crafter studio using utf encoding to copy filtered resources copying resources maven compiler plugin testcompile default testcompile crafter studio changes detected recompiling the module compiling source files to home jose github craftercms src studio target test classes compilation error home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method getdeploymenthistory java lang string java lang object java lang object java lang object java lang object java lang object int location variable local of type org craftercms studio impl repository git gitcontentrepository errors build failure total time s finished at failed to execute goal org apache maven plugins maven compiler plugin testcompile default testcompile on project crafter studio compilation failure compilation failure home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method getdeploymenthistory java lang string java lang object java lang object java lang object java lang object java lang object int location variable local of type org craftercms studio impl repository git gitcontentrepository to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles specs version studio develop branch studio support x branch os linux browser n a additional context n a
0
1,616
17,606,624,051
IssuesEvent
2021-08-17 17:57:58
timberio/vector
https://api.github.com/repos/timberio/vector
opened
Running out of disk space should not stall disk buffering.
type: bug domain: buffers domain: reliability
Presently, if running a configuration with disk buffers enabled, the process of running out of space will result in a panic being thrown in `lib/vector-core/buffers/src/disk/leveldb_buffer/writer.rs:189` due to unwrapping a result which ends up being the I/O error and not the expected success of the write. Since disk buffering will correctly stop buffering once we hit the configured `buffer.max_size`, we should consider running out of disk space to be approximately the same condition. This might be something we can only generally capture, given that I don't know if there's a way, at the type level, to assert that an I/O error is specifically related to running out of disk space. We'd be doing a best effort check on the error's description. One might also argue that an I/O error in disk buffering on the write path should simply continue along, applying normal backpressure, but that's a bigger discussion.
True
Running out of disk space should not stall disk buffering. - Presently, if running a configuration with disk buffers enabled, the process of running out of space will result in a panic being thrown in `lib/vector-core/buffers/src/disk/leveldb_buffer/writer.rs:189` due to unwrapping a result which ends up being the I/O error and not the expected success of the write. Since disk buffering will correctly stop buffering once we hit the configured `buffer.max_size`, we should consider running out of disk space to be approximately the same condition. This might be something we can only generally capture, given that I don't know if there's a way, at the type level, to assert that an I/O error is specifically related to running out of disk space. We'd be doing a best effort check on the error's description. One might also argue that an I/O error in disk buffering on the write path should simply continue along, applying normal backpressure, but that's a bigger discussion.
reli
running out of disk space should not stall disk buffering presently if running a configuration with disk buffers enabled the process of running out of space will result in a panic being thrown in lib vector core buffers src disk leveldb buffer writer rs due to unwrapping a result which ends up being the i o error and not the expected success of the write since disk buffering will correctly stop buffering once we hit the configured buffer max size we should consider running out of disk space to be approximately the same condition this might be something we can only generally capture given that i don t know if there s a way at the type level to assert that an i o error is specifically related to running out of disk space we d be doing a best effort check on the error s description one might also argue that an i o error in disk buffering on the write path should simply continue along applying normal backpressure but that s a bigger discussion
1
578,914
17,156,555,898
IssuesEvent
2021-07-14 07:42:01
vrchatapi/specification
https://api.github.com/repos/vrchatapi/specification
closed
Issue with accountDeletionDate in CurrentUser.yaml
Priority: Medium Status: Accepted Type: Maintenance
``` accountDeletionDate: oneOf: - $ref: ./Date.yaml - nullable: true ``` I think something is wrong with this piece but I don't know enough about the OpenAPI specification to fix it. Here are some links I've found regarding this: https://github.com/Redocly/redoc/issues/1303 https://stackoverflow.com/questions/40920441/how-to-specify-a-property-can-be-null-or-a-reference-with-swagger I tried doing some of the things in those links but I ended up just making it worse
1.0
Issue with accountDeletionDate in CurrentUser.yaml - ``` accountDeletionDate: oneOf: - $ref: ./Date.yaml - nullable: true ``` I think something is wrong with this piece but I don't know enough about the OpenAPI specification to fix it. Here are some links I've found regarding this: https://github.com/Redocly/redoc/issues/1303 https://stackoverflow.com/questions/40920441/how-to-specify-a-property-can-be-null-or-a-reference-with-swagger I tried doing some of the things in those links but I ended up just making it worse
non_reli
issue with accountdeletiondate in currentuser yaml accountdeletiondate oneof ref date yaml nullable true i think something is wrong with this piece but i don t know enough about the openapi specification to fix it here are some links i ve found regarding this i tried doing some of the things in those links but i ended up just making it worse
0
3,033
31,778,268,082
IssuesEvent
2023-09-12 15:38:38
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] SupportSplitAndRetry for GpuRangeExec
feature request reliability
**Is your feature request related to a problem? Please describe.** GpuRangeExec is mostly done on the CPU. Only once it has a start, end, and step for a batch does it move onto the GPU for processing. The main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch. For split and retry we would need to make sure that if we see a split that we can drop the number of `rowsThisBatch`, but not too small.
True
[FEA] SupportSplitAndRetry for GpuRangeExec - **Is your feature request related to a problem? Please describe.** GpuRangeExec is mostly done on the CPU. Only once it has a start, end, and step for a batch does it move onto the GPU for processing. The main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch. For split and retry we would need to make sure that if we see a split that we can drop the number of `rowsThisBatch`, but not too small.
reli
supportsplitandretry for gpurangeexec is your feature request related to a problem please describe gpurangeexec is mostly done on the cpu only once it has a start end and step for a batch does it move onto the gpu for processing the main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch for split and retry we would need to make sure that if we see a split that we can drop the number of rowsthisbatch but not too small
1
576
8,689,972,313
IssuesEvent
2018-12-03 20:12:08
rook/rook
https://api.github.com/repos/rook/rook
closed
Proper cleanup after lost nodes/OSDs
operator reliability wontfix
* Bug Report **Deviation from expected behavior:** Configmaps and deployments related to terminated k8s nodes are left dangling, and OSDs are not taken out of the ceph cluster. **Expected behavior:** I would expect that configmaps and deployments for deleted/terminated/missing nodes to be removed by the operator. **How to reproduce it (minimal and precise):** Create a multi node K8s cluster, and add a OSDs on each node. Terminate/delete one of the k8s nodes and watch that no cleanup is not performed. To cleanup you have to 1. Delete the job and configmap in the rook-ceph namespace related the dead/removed node 2. Delete the deployment in the rook-ceph namespace related to the osd that was present on the dead/removed node 3. Delete the two configmaps related to the dead/removed node in the rook-ceph-system namespace 4. Restart the rook operator. Also dead OSDs should be removed from the ceph cluster by issuing `ceph osd purge osd.ID --yes-i-really-mean-it`, where ID is the OSD that was previously running on the terminated node. Refer to this thread on slack for more details https://rook-io.slack.com/messages/C46Q5UC05/convo/C46Q5UC05-1534330458.000100/ **Environment**: * OS (e.g. from /etc/os-release): CoreOS * Cloud provider or hardware configuration: AWS * Rook version (use `rook version` inside of a Rook Pod): 0.8.1 * Kubernetes version (use `kubectl version`): 1.9.3 * Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift): On AWS, built with kube-aws * Storage backend status (e.g. for Ceph use `ceph health` in the [Rook Ceph toolbox](https://rook.io/docs/Rook/master/toolbox.html)): OK
True
Proper cleanup after lost nodes/OSDs - * Bug Report **Deviation from expected behavior:** Configmaps and deployments related to terminated k8s nodes are left dangling, and OSDs are not taken out of the ceph cluster. **Expected behavior:** I would expect that configmaps and deployments for deleted/terminated/missing nodes to be removed by the operator. **How to reproduce it (minimal and precise):** Create a multi node K8s cluster, and add a OSDs on each node. Terminate/delete one of the k8s nodes and watch that no cleanup is not performed. To cleanup you have to 1. Delete the job and configmap in the rook-ceph namespace related the dead/removed node 2. Delete the deployment in the rook-ceph namespace related to the osd that was present on the dead/removed node 3. Delete the two configmaps related to the dead/removed node in the rook-ceph-system namespace 4. Restart the rook operator. Also dead OSDs should be removed from the ceph cluster by issuing `ceph osd purge osd.ID --yes-i-really-mean-it`, where ID is the OSD that was previously running on the terminated node. Refer to this thread on slack for more details https://rook-io.slack.com/messages/C46Q5UC05/convo/C46Q5UC05-1534330458.000100/ **Environment**: * OS (e.g. from /etc/os-release): CoreOS * Cloud provider or hardware configuration: AWS * Rook version (use `rook version` inside of a Rook Pod): 0.8.1 * Kubernetes version (use `kubectl version`): 1.9.3 * Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift): On AWS, built with kube-aws * Storage backend status (e.g. for Ceph use `ceph health` in the [Rook Ceph toolbox](https://rook.io/docs/Rook/master/toolbox.html)): OK
reli
proper cleanup after lost nodes osds bug report deviation from expected behavior configmaps and deployments related to terminated nodes are left dangling and osds are not taken out of the ceph cluster expected behavior i would expect that configmaps and deployments for deleted terminated missing nodes to be removed by the operator how to reproduce it minimal and precise create a multi node cluster and add a osds on each node terminate delete one of the nodes and watch that no cleanup is not performed to cleanup you have to delete the job and configmap in the rook ceph namespace related the dead removed node delete the deployment in the rook ceph namespace related to the osd that was present on the dead removed node delete the two configmaps related to the dead removed node in the rook ceph system namespace restart the rook operator also dead osds should be removed from the ceph cluster by issuing ceph osd purge osd id yes i really mean it where id is the osd that was previously running on the terminated node refer to this thread on slack for more details environment os e g from etc os release coreos cloud provider or hardware configuration aws rook version use rook version inside of a rook pod kubernetes version use kubectl version kubernetes cluster type e g tectonic gke openshift on aws built with kube aws storage backend status e g for ceph use ceph health in the ok
1
757
10,476,136,405
IssuesEvent
2019-09-23 17:54:25
microsoft/BotFramework-DirectLineJS
https://api.github.com/repos/microsoft/BotFramework-DirectLineJS
opened
Happy path: with conversation ID, typing activity can be sent
Reliability Streaming Extensions
1. Start a conversation by specifying a conversation ID 1. Send a typing activity 1. Make sure the bot can receive it
True
Happy path: with conversation ID, typing activity can be sent - 1. Start a conversation by specifying a conversation ID 1. Send a typing activity 1. Make sure the bot can receive it
reli
happy path with conversation id typing activity can be sent start a conversation by specifying a conversation id send a typing activity make sure the bot can receive it
1
2,328
24,711,085,152
IssuesEvent
2022-10-20 00:51:59
ppy/osu
https://api.github.com/repos/ppy/osu
closed
Game crashing after map loaded from the editor.
type:online type:reliability
### Type Crash to desktop ### Bug description I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one. The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing. 5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black. **The game crashed** after like **3 seconds**. This stopped happening after I wrote this and I still don't know what caused this reactive behavior. ### Screenshots or videos I don't have any photos/videos that show this in action. ### Version osu! 2022.1008.2-lazer ### Logs I'm going to send them all, but I found that runtime is most interesting... [database.log](https://github.com/ppy/osu/files/9785324/database.log) [input.log](https://github.com/ppy/osu/files/9785326/input.log) [network.log](https://github.com/ppy/osu/files/9785327/network.log) [performance.log](https://github.com/ppy/osu/files/9785328/performance.log) [runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log) [updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
True
Game crashing after map loaded from the editor. - ### Type Crash to desktop ### Bug description I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one. The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing. 5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black. **The game crashed** after like **3 seconds**. This stopped happening after I wrote this and I still don't know what caused this reactive behavior. ### Screenshots or videos I don't have any photos/videos that show this in action. ### Version osu! 2022.1008.2-lazer ### Logs I'm going to send them all, but I found that runtime is most interesting... [database.log](https://github.com/ppy/osu/files/9785324/database.log) [input.log](https://github.com/ppy/osu/files/9785326/input.log) [network.log](https://github.com/ppy/osu/files/9785327/network.log) [performance.log](https://github.com/ppy/osu/files/9785328/performance.log) [runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log) [updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
reli
game crashing after map loaded from the editor type crash to desktop bug description i was using osu without an internet connection and wanted to play some maps i got to the song selector and loaded a map but it was taking a lot of time so i canceled the map and reloaded another one the same thing happened so i loaded another map in the editor to see if there was something wrong but there wasn t so i played the map from the editor this time everything was going well but after seconds of playing the game froze but the audio was still playing seconds later the screen went black but the audio was still playing seconds after that the audio stopped playing and the usual osu lazer has stopped working popped up but i was waiting then the game played audio again but the screen was still black the game crashed after like seconds this stopped happening after i wrote this and i still don t know what caused this reactive behavior screenshots or videos i don t have any photos videos that show this in action version osu lazer logs i m going to send them all but i found that runtime is most interesting
1
237,972
7,768,754,195
IssuesEvent
2018-06-03 21:27:48
TerraFusion/basicFusion
https://api.github.com/repos/TerraFusion/basicFusion
closed
Modify generation scripts for BF output directories
Medium Priority enhancement
Modify the generation scripts so that the BF output files are stored in some sort of directory hierarchy based on dates.
1.0
Modify generation scripts for BF output directories - Modify the generation scripts so that the BF output files are stored in some sort of directory hierarchy based on dates.
non_reli
modify generation scripts for bf output directories modify the generation scripts so that the bf output files are stored in some sort of directory hierarchy based on dates
0
317
6,558,869,926
IssuesEvent
2017-09-06 23:48:32
waggle-sensor/beehive-server
https://api.github.com/repos/waggle-sensor/beehive-server
opened
Prototype static version of beehive
reliability
One _possible_ improvement we can do is build a static version of beehive which is regenerated on a schedule. This would dramatically improve page serving performance across the board. This also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up. I think this is still worth prototyping, even though we now have nginx performing caching and have moved off the development server. As an example, the build-index tool in the data-exporter generates a "friendly" summary of all the datasets to make sure things look reasonable.
True
Prototype static version of beehive - One _possible_ improvement we can do is build a static version of beehive which is regenerated on a schedule. This would dramatically improve page serving performance across the board. This also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up. I think this is still worth prototyping, even though we now have nginx performing caching and have moved off the development server. As an example, the build-index tool in the data-exporter generates a "friendly" summary of all the datasets to make sure things look reasonable.
reli
prototype static version of beehive one possible improvement we can do is build a static version of beehive which is regenerated on a schedule this would dramatically improve page serving performance across the board this also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up i think this is still worth prototyping even though we now have nginx performing caching and have moved off the development server as an example the build index tool in the data exporter generates a friendly summary of all the datasets to make sure things look reasonable
1
13,067
8,788,279,540
IssuesEvent
2018-12-20 21:36:13
istio/istio
https://api.github.com/repos/istio/istio
closed
Citadel in CrashLoopBackOff
area/security stale
**Describe the bug** Use helm to create a yaml file including grafana **Expected behavior** running grafana container **Steps to reproduce the bug** helm template ../istio/install/kubernetes/helm/istio --name istio --namespace istio-system --set grafana.enabled=true --set grafana.security.enabled=true --set grafana.security.adminUser=admin --set grafana.security.adminPassword=mypass > istio.yaml kubectl create ns istio-system kubectl apply -n istio-system -f istio.yaml ``` $ kubectl get all --all-namespaces | grep -i istio istio-system pod/grafana-67c5b957f9-l79mq 0/1 CreateContainerConfigError 0 8m istio-system pod/istio-citadel-7d9c98786d-prpwn 0/1 CrashLoopBackOff 6 8m istio-system pod/istio-cleanup-secrets-s5fhb 0/1 Completed 0 8m istio-system pod/istio-egressgateway-7bdd8c695f-t6g6f 1/1 Running 0 8m istio-system pod/istio-galley-6c799667d4-jscww 0/1 ContainerCreating 0 8m ... istio-system pod/istio-sidecar-injector-c8c4c568b-55h7z 0/1 ContainerCreating 0 8m ... ``` **Version** ``` $ istioctl version Version: 1.0.1 GitRevision: 42773aacced474d97159902d20579a25b1f98106 User: root@832d5020b1d4 Hub: gcr.io/istio-release GolangVersion: go1.10.1 BuildStatus: Clean $ kubectl version Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.5", GitCommit:"32ac1c9073b132b8ba18aa830f46b77dcceb0723", GitTreeState:"clean", BuildDate:"2018-06-21T11:34:22Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"} ``` **Is Istio Auth enabled or not?** No idea **Environment** Kops w/coreOS on AWS **Cluster state** [istio-dump.tar.gz](https://github.com/istio/istio/files/2358972/istio-dump.tar.gz)
True
Citadel in CrashLoopBackOff - **Describe the bug** Use helm to create a yaml file including grafana **Expected behavior** running grafana container **Steps to reproduce the bug** helm template ../istio/install/kubernetes/helm/istio --name istio --namespace istio-system --set grafana.enabled=true --set grafana.security.enabled=true --set grafana.security.adminUser=admin --set grafana.security.adminPassword=mypass > istio.yaml kubectl create ns istio-system kubectl apply -n istio-system -f istio.yaml ``` $ kubectl get all --all-namespaces | grep -i istio istio-system pod/grafana-67c5b957f9-l79mq 0/1 CreateContainerConfigError 0 8m istio-system pod/istio-citadel-7d9c98786d-prpwn 0/1 CrashLoopBackOff 6 8m istio-system pod/istio-cleanup-secrets-s5fhb 0/1 Completed 0 8m istio-system pod/istio-egressgateway-7bdd8c695f-t6g6f 1/1 Running 0 8m istio-system pod/istio-galley-6c799667d4-jscww 0/1 ContainerCreating 0 8m ... istio-system pod/istio-sidecar-injector-c8c4c568b-55h7z 0/1 ContainerCreating 0 8m ... ``` **Version** ``` $ istioctl version Version: 1.0.1 GitRevision: 42773aacced474d97159902d20579a25b1f98106 User: root@832d5020b1d4 Hub: gcr.io/istio-release GolangVersion: go1.10.1 BuildStatus: Clean $ kubectl version Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.5", GitCommit:"32ac1c9073b132b8ba18aa830f46b77dcceb0723", GitTreeState:"clean", BuildDate:"2018-06-21T11:34:22Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"} ``` **Is Istio Auth enabled or not?** No idea **Environment** Kops w/coreOS on AWS **Cluster state** [istio-dump.tar.gz](https://github.com/istio/istio/files/2358972/istio-dump.tar.gz)
non_reli
citadel in crashloopbackoff describe the bug use helm to create a yaml file including grafana expected behavior running grafana container steps to reproduce the bug helm template istio install kubernetes helm istio name istio namespace istio system set grafana enabled true set grafana security enabled true set grafana security adminuser admin set grafana security adminpassword mypass istio yaml kubectl create ns istio system kubectl apply n istio system f istio yaml kubectl get all all namespaces grep i istio istio system pod grafana createcontainerconfigerror istio system pod istio citadel prpwn crashloopbackoff istio system pod istio cleanup secrets completed istio system pod istio egressgateway running istio system pod istio galley jscww containercreating istio system pod istio sidecar injector containercreating version istioctl version version gitrevision user root hub gcr io istio release golangversion buildstatus clean kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux is istio auth enabled or not no idea environment kops w coreos on aws cluster state
0
3,023
31,623,332,118
IssuesEvent
2023-09-06 02:02:13
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] Add retry support to GpuSorter.fullySortBatch
feature request reliability
**Is your feature request related to a problem? Please describe.** This is used both for GpuTopN and GpuSortEachBatchIterator.
True
[FEA] Add retry support to GpuSorter.fullySortBatch - **Is your feature request related to a problem? Please describe.** This is used both for GpuTopN and GpuSortEachBatchIterator.
reli
add retry support to gpusorter fullysortbatch is your feature request related to a problem please describe this is used both for gputopn and gpusorteachbatchiterator
1
2,238
24,364,117,445
IssuesEvent
2022-10-03 13:58:39
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] Allow broadcast tables to be spilled if necessary
P1 reliability
**Is your feature request related to a problem? Please describe.** Currently broadcast tables are "intentionally leaked" in GPU memory, as they are only cleaned up when garbage collected. It would be nice if instead of leaving them permanently in GPU memory until garbage collected we were able to spill them to host memory (and ultimately disk if necessary) when GPU memory is low. **Describe the solution you'd like** We should add broadcast tables to the spillable buffer framework. There could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used (as the spill framework currently doesn't migrate a buffer's recorded location from host back to device once spilled). However running a bit slow beats crashing due to OOM, so this would be a good first step.
True
[FEA] Allow broadcast tables to be spilled if necessary - **Is your feature request related to a problem? Please describe.** Currently broadcast tables are "intentionally leaked" in GPU memory, as they are only cleaned up when garbage collected. It would be nice if instead of leaving them permanently in GPU memory until garbage collected we were able to spill them to host memory (and ultimately disk if necessary) when GPU memory is low. **Describe the solution you'd like** We should add broadcast tables to the spillable buffer framework. There could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used (as the spill framework currently doesn't migrate a buffer's recorded location from host back to device once spilled). However running a bit slow beats crashing due to OOM, so this would be a good first step.
reli
allow broadcast tables to be spilled if necessary is your feature request related to a problem please describe currently broadcast tables are intentionally leaked in gpu memory as they are only cleaned up when garbage collected it would be nice if instead of leaving them permanently in gpu memory until garbage collected we were able to spill them to host memory and ultimately disk if necessary when gpu memory is low describe the solution you d like we should add broadcast tables to the spillable buffer framework there could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used as the spill framework currently doesn t migrate a buffer s recorded location from host back to device once spilled however running a bit slow beats crashing due to oom so this would be a good first step
1
209,550
16,039,599,602
IssuesEvent
2021-04-22 05:46:22
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] AutoFollowIT testDataStreams_autoFollowAfterDataStreamCreated failing
:Core/Features/Data streams :Distributed/CCR >test-failure
Subsequent tests failed as well, but I suspect they are due to this initial failure. ``` {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400} ``` **Build scan:** https://gradle-enterprise.elastic.co/s/q4m7uwwqj7mei/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testDataStreams_autoFollowAfterDataStreamCreated **Reproduction line:** `./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated" -Dtests.seed=197AA5A92FE81C06 -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Argentina/San_Juan -Druntime.java=8` **Applicable branches:** 7.x **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testDataStreams_autoFollowAfterDataStreamCreated **Failure excerpt:** ``` org.elasticsearch.client.ResponseException: method [PUT], host [http://[::1]:45566], URI [/.ds-logs-syslog-prod-2021.04.22-000001/_ccr/follow?wait_for_active_shards=1], status line [HTTP/1.1 400 Bad Request] {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400} at __randomizedtesting.SeedInfo.seed([197AA5A92FE81C06:AD357CF326BDC165]:0) at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:109) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:81) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:77) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:73) at org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated(AutoFollowIT.java:319) at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) ```
1.0
[CI] AutoFollowIT testDataStreams_autoFollowAfterDataStreamCreated failing - Subsequent tests failed as well, but I suspect they are due to this initial failure. ``` {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400} ``` **Build scan:** https://gradle-enterprise.elastic.co/s/q4m7uwwqj7mei/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testDataStreams_autoFollowAfterDataStreamCreated **Reproduction line:** `./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated" -Dtests.seed=197AA5A92FE81C06 -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Argentina/San_Juan -Druntime.java=8` **Applicable branches:** 7.x **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testDataStreams_autoFollowAfterDataStreamCreated **Failure excerpt:** ``` org.elasticsearch.client.ResponseException: method [PUT], host [http://[::1]:45566], URI [/.ds-logs-syslog-prod-2021.04.22-000001/_ccr/follow?wait_for_active_shards=1], status line [HTTP/1.1 400 Bad Request] {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400} at __randomizedtesting.SeedInfo.seed([197AA5A92FE81C06:AD357CF326BDC165]:0) at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:109) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:81) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:77) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:73) at org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated(AutoFollowIT.java:319) at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) ```
non_reli
autofollowit testdatastreams autofollowafterdatastreamcreated failing subsequent tests failed as well but i suspect they are due to this initial failure error root cause and follower index settings must be identical type illegal argument exception reason the leader index settings and follower index settings must be identical status build scan reproduction line gradlew x pack plugin ccr qa multi cluster follow cluster tests org elasticsearch xpack ccr autofollowit testdatastreams autofollowafterdatastreamcreated dtests seed dtests locale sr latn ba dtests timezone america argentina san juan druntime java applicable branches x reproduces locally no failure history failure excerpt org elasticsearch client responseexception method host uri status line error root cause and follower index settings must be identical type illegal argument exception reason the leader index settings and follower index settings must be identical status at randomizedtesting seedinfo seed at org elasticsearch client restclient convertresponse restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr autofollowit testdatastreams autofollowafterdatastreamcreated autofollowit java at sun reflect nativemethodaccessorimpl nativemethodaccessorimpl java at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at java lang thread run thread java
0
199,924
6,996,156,657
IssuesEvent
2017-12-15 22:41:47
GoogleCloudPlatform/google-cloud-go
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-go
closed
logging: automatically detect GKE and GAE Flex resources
api: logging help wanted priority: p2 type: feature request
The creation of a new Logger can automatically set ressources for GCE instances, but it cannot do it yet for GKE Containers. Having type:gce_instance can be confusing for code running inside a container. You end up with logs from the containers appear as if the hosting machine was producing them. Proposed behaviour: * When code runs inside a Google Container Engine * the type is set to 'container', * the labels: pod_id, cluster_name, container_name, namespace_id, are populated as well.
1.0
logging: automatically detect GKE and GAE Flex resources - The creation of a new Logger can automatically set ressources for GCE instances, but it cannot do it yet for GKE Containers. Having type:gce_instance can be confusing for code running inside a container. You end up with logs from the containers appear as if the hosting machine was producing them. Proposed behaviour: * When code runs inside a Google Container Engine * the type is set to 'container', * the labels: pod_id, cluster_name, container_name, namespace_id, are populated as well.
non_reli
logging automatically detect gke and gae flex resources the creation of a new logger can automatically set ressources for gce instances but it cannot do it yet for gke containers having type gce instance can be confusing for code running inside a container you end up with logs from the containers appear as if the hosting machine was producing them proposed behaviour when code runs inside a google container engine the type is set to container the labels pod id cluster name container name namespace id are populated as well
0
1,598
17,376,274,704
IssuesEvent
2021-07-30 21:56:58
clowdr-app/clowdr
https://api.github.com/repos/clowdr-app/clowdr
closed
Automatically clean up phantom room participants
A-rooms C-bug D-medium S-actions T-reliability U-soon
If you join the room, there is none there, but they are still listed as active <img width="494" alt="Screen Shot 2021-01-21 at 5 31 26 PM" src="https://user-images.githubusercontent.com/506426/105420581-9f7d0400-5c0e-11eb-8e36-07e627d21c63.png">
True
Automatically clean up phantom room participants - If you join the room, there is none there, but they are still listed as active <img width="494" alt="Screen Shot 2021-01-21 at 5 31 26 PM" src="https://user-images.githubusercontent.com/506426/105420581-9f7d0400-5c0e-11eb-8e36-07e627d21c63.png">
reli
automatically clean up phantom room participants if you join the room there is none there but they are still listed as active img width alt screen shot at pm src
1
401
7,308,657,544
IssuesEvent
2018-02-28 09:10:36
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Crash in SqlConnection.Close() when using TransactionScope on linux.
area-System.Data.SqlClient os-linux tenet-reliability
I'm using SqlClient 4.5.0-preview1-26207-04 in netcoreapp2.0. ``` .NET Command Line Tools (2.0.0) Product Information: Version: 2.0.0 Commit SHA-1 hash: cdcd1928c9 Runtime Environment: OS Name: centos OS Version: 7 OS Platform: Linux RID: centos.7-x64 Base Path: /usr/share/dotnet/sdk/2.0.0/ Microsoft .NET Core Shared Framework Host Version : 2.0.0 Build : e8b8861ac7faf042c87a5c2f9f2d04c98b69f28d ``` The following snippet would crash on the above machine (removed some code for brevity). Works without issues on Windows. ```c# using (var txscope = new TransactionScope(TransactionScopeAsyncFlowOption.Enabled)) using (var cn = _dBProvider.GetWriteConnection()) { await cn.OpenAsync(); if (cmd.RecordCertificationStatus) { await cn.ExecuteAsync(@"INSERT INTO (...) (@iduser, @idtitle, @value, getdate()); ", new { @iduser = cmd.SubId, @idtitle=cmd.EntityId, @value = cmd.Model.IsCertified }); } await cn.ExecuteAsync(sql, data); txscope.Complete(); } ``` Both statements are inserts. Stacktrace is below: ``` "System.NullReferenceException: Object reference not set to an instance of an object. at System.Data.ProviderBase.DbConnectionPool.DeactivateObject(DbConnectionInternal obj) at System.Data.ProviderBase.DbConnectionPool.PutObject(DbConnectionInternal obj, Object owningObject) at System.Data.ProviderBase.DbConnectionInternal.CloseConnection(DbConnection owningObject, DbConnectionFactory connectionFactory) at System.Data.SqlClient.SqlConnection.Close() at System.Data.SqlClient.SqlConnection.Dispose(Boolean disposing) at System.ComponentModel.Component.Dispose() at ...TitleUpdate.Handler.<HandleCore>d__2.MoveNext() in ...TitleUpdate.cs:line 189 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at MediatR.Internal.RequestHandlerWrapperImpl`1.<>c__DisplayClass0_0.<<Handle>g__Handler0>d.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult() at ...ValidationBehavior`2.<Handle>d__2.MoveNext() in ...ValidationBehavior.cs:line 34 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at ....TitlesController.<Update>d__5.MoveNext() in TitleController.cs:line 98 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeActionMethodAsync>d__12.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeNextActionFilterAsync>d__10.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Rethrow(ActionExecutedContext context) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Next(State& next, Scope& scope, Object& state, Boolean& isCompleted) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeInnerFilterAsync>d__14.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ResourceInvoker.<InvokeNextExceptionFilterAsync>d__23.MoveNext()" ```
True
Crash in SqlConnection.Close() when using TransactionScope on linux. - I'm using SqlClient 4.5.0-preview1-26207-04 in netcoreapp2.0. ``` .NET Command Line Tools (2.0.0) Product Information: Version: 2.0.0 Commit SHA-1 hash: cdcd1928c9 Runtime Environment: OS Name: centos OS Version: 7 OS Platform: Linux RID: centos.7-x64 Base Path: /usr/share/dotnet/sdk/2.0.0/ Microsoft .NET Core Shared Framework Host Version : 2.0.0 Build : e8b8861ac7faf042c87a5c2f9f2d04c98b69f28d ``` The following snippet would crash on the above machine (removed some code for brevity). Works without issues on Windows. ```c# using (var txscope = new TransactionScope(TransactionScopeAsyncFlowOption.Enabled)) using (var cn = _dBProvider.GetWriteConnection()) { await cn.OpenAsync(); if (cmd.RecordCertificationStatus) { await cn.ExecuteAsync(@"INSERT INTO (...) (@iduser, @idtitle, @value, getdate()); ", new { @iduser = cmd.SubId, @idtitle=cmd.EntityId, @value = cmd.Model.IsCertified }); } await cn.ExecuteAsync(sql, data); txscope.Complete(); } ``` Both statements are inserts. Stacktrace is below: ``` "System.NullReferenceException: Object reference not set to an instance of an object. at System.Data.ProviderBase.DbConnectionPool.DeactivateObject(DbConnectionInternal obj) at System.Data.ProviderBase.DbConnectionPool.PutObject(DbConnectionInternal obj, Object owningObject) at System.Data.ProviderBase.DbConnectionInternal.CloseConnection(DbConnection owningObject, DbConnectionFactory connectionFactory) at System.Data.SqlClient.SqlConnection.Close() at System.Data.SqlClient.SqlConnection.Dispose(Boolean disposing) at System.ComponentModel.Component.Dispose() at ...TitleUpdate.Handler.<HandleCore>d__2.MoveNext() in ...TitleUpdate.cs:line 189 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at MediatR.Internal.RequestHandlerWrapperImpl`1.<>c__DisplayClass0_0.<<Handle>g__Handler0>d.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult() at ...ValidationBehavior`2.<Handle>d__2.MoveNext() in ...ValidationBehavior.cs:line 34 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at ....TitlesController.<Update>d__5.MoveNext() in TitleController.cs:line 98 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeActionMethodAsync>d__12.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeNextActionFilterAsync>d__10.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Rethrow(ActionExecutedContext context) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Next(State& next, Scope& scope, Object& state, Boolean& isCompleted) at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeInnerFilterAsync>d__14.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.AspNetCore.Mvc.Internal.ResourceInvoker.<InvokeNextExceptionFilterAsync>d__23.MoveNext()" ```
reli
crash in sqlconnection close when using transactionscope on linux i m using sqlclient in net command line tools product information version commit sha hash runtime environment os name centos os version os platform linux rid centos base path usr share dotnet sdk microsoft net core shared framework host version build the following snippet would crash on the above machine removed some code for brevity works without issues on windows c using var txscope new transactionscope transactionscopeasyncflowoption enabled using var cn dbprovider getwriteconnection await cn openasync if cmd recordcertificationstatus await cn executeasync insert into iduser idtitle value getdate new iduser cmd subid idtitle cmd entityid value cmd model iscertified await cn executeasync sql data txscope complete both statements are inserts stacktrace is below system nullreferenceexception object reference not set to an instance of an object at system data providerbase dbconnectionpool deactivateobject dbconnectioninternal obj at system data providerbase dbconnectionpool putobject dbconnectioninternal obj object owningobject at system data providerbase dbconnectioninternal closeconnection dbconnection owningobject dbconnectionfactory connectionfactory at system data sqlclient sqlconnection close at system data sqlclient sqlconnection dispose boolean disposing at system componentmodel component dispose at titleupdate handler d movenext in titleupdate cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at mediatr internal requesthandlerwrapperimpl c g d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at validationbehavior d movenext in validationbehavior cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at titlescontroller d movenext in titlecontroller cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft aspnetcore mvc internal controlleractioninvoker rethrow actionexecutedcontext context at microsoft aspnetcore mvc internal controlleractioninvoker next state next scope scope object state boolean iscompleted at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal resourceinvoker d movenext
1
383,407
11,355,798,045
IssuesEvent
2020-01-24 20:56:43
yalla-coop/presspad
https://api.github.com/repos/yalla-coop/presspad
opened
Create updates for intern, hosts, and organisations
priority-5
- [ ] When intern request a stay - [ ] host (`stayRequest`) - [ ] org (`stayRequest`) - [ ] when intern accept a request - [ ] intern (`stayApproved`) - [ ] org (`stayApproved`) - [ ] when host reject a request - [ ] intern (`stayRejected`) - [ ] org (`stayRejected`) - [ ] when stay completed - [ ] intern (`giveReviewReminder`) - [ ] host (`giveReviewReminder`) - [ ] org (`stayCompleted`) - [ ] when host give an intern a review - [ ] intern (`getReview`) - [ ] org (`getReview`) - [ ] when intern give a host a review - [ ] host (`getReview`) - [ ] org (`internGaveHostReview`)
1.0
Create updates for intern, hosts, and organisations - - [ ] When intern request a stay - [ ] host (`stayRequest`) - [ ] org (`stayRequest`) - [ ] when intern accept a request - [ ] intern (`stayApproved`) - [ ] org (`stayApproved`) - [ ] when host reject a request - [ ] intern (`stayRejected`) - [ ] org (`stayRejected`) - [ ] when stay completed - [ ] intern (`giveReviewReminder`) - [ ] host (`giveReviewReminder`) - [ ] org (`stayCompleted`) - [ ] when host give an intern a review - [ ] intern (`getReview`) - [ ] org (`getReview`) - [ ] when intern give a host a review - [ ] host (`getReview`) - [ ] org (`internGaveHostReview`)
non_reli
create updates for intern hosts and organisations when intern request a stay host stayrequest org stayrequest when intern accept a request intern stayapproved org stayapproved when host reject a request intern stayrejected org stayrejected when stay completed intern givereviewreminder host givereviewreminder org staycompleted when host give an intern a review intern getreview org getreview when intern give a host a review host getreview org interngavehostreview
0
35,742
5,005,664,868
IssuesEvent
2016-12-12 11:28:49
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster
kind/upgrade-test-failure non-release-blocker
Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4) Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
1.0
[Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster - Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4) Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
non_reli
scheduledjob should replace jobs when replaceconcurrent in ci kubernetes gce upgrade cluster test has been failing in sample failure based on spreadsheet tracking upgrade test failures created by krousey
0
152,864
19,697,885,119
IssuesEvent
2022-01-12 14:00:30
rarkins/spdx-maven-plugin-demo
https://api.github.com/repos/rarkins/spdx-maven-plugin-demo
opened
CVE-2020-35490 (High) detected in jackson-databind-2.9.10.4.jar
security vulnerability
## CVE-2020-35490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p> <p> Dependency Hierarchy: - spdx-tools-2.2.1.jar (Root Library) - :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35490 (High) detected in jackson-databind-2.9.10.4.jar - ## CVE-2020-35490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p> <p> Dependency Hierarchy: - spdx-tools-2.2.1.jar (Root Library) - :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_reli
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spdx tools jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
1,972
22,296,923,997
IssuesEvent
2022-06-13 03:27:02
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] Stop running task attempts on executors that encounter "sticky" CUDA errors
P1 reliability
**Is your feature request related to a problem? Please describe.** Certain CUDA errors, like illegal memory access, are "sticky," meaning that all CUDA operations to the GPU after the error will continue to return the same error over and over. No GPU operations will succeed after that point. **Describe the solution you'd like** The RAPIDS Accelerator should take measures to prevent further task execution on the executor once these "sticky" exceptions are detected. Tearing down the executor process is probably the best option, at least in the short-term. Without an external shuffle handler we will lose the shuffle of tasks that have completed, but this is probably a better way to "fail fast" then allow the executor to keep accepting new tasks only to have them fail the first time they touch the GPU.
True
[FEA] Stop running task attempts on executors that encounter "sticky" CUDA errors - **Is your feature request related to a problem? Please describe.** Certain CUDA errors, like illegal memory access, are "sticky," meaning that all CUDA operations to the GPU after the error will continue to return the same error over and over. No GPU operations will succeed after that point. **Describe the solution you'd like** The RAPIDS Accelerator should take measures to prevent further task execution on the executor once these "sticky" exceptions are detected. Tearing down the executor process is probably the best option, at least in the short-term. Without an external shuffle handler we will lose the shuffle of tasks that have completed, but this is probably a better way to "fail fast" then allow the executor to keep accepting new tasks only to have them fail the first time they touch the GPU.
reli
stop running task attempts on executors that encounter sticky cuda errors is your feature request related to a problem please describe certain cuda errors like illegal memory access are sticky meaning that all cuda operations to the gpu after the error will continue to return the same error over and over no gpu operations will succeed after that point describe the solution you d like the rapids accelerator should take measures to prevent further task execution on the executor once these sticky exceptions are detected tearing down the executor process is probably the best option at least in the short term without an external shuffle handler we will lose the shuffle of tasks that have completed but this is probably a better way to fail fast then allow the executor to keep accepting new tasks only to have them fail the first time they touch the gpu
1
2,229
24,319,431,624
IssuesEvent
2022-09-30 09:23:48
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
DefaultActorClock is not thread safe
kind/bug area/performance severity/high severity/mid area/reliability
**Describe the bug** It seems we now create the ActorScheduler with a predefined actor clock, either ControlledClock or the DefaultActorClock https://github.com/camunda/zeebe/blob/main/dist/src/main/java/io/camunda/zeebe/shared/ActorClockConfiguration.java#L34-L35 The problem is that the DefaultActorClock is not thread safe! Previously we created the ActorScheduler without clock, which caused [that each thread has its own clock](https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L64). Then there is no issue. If we use the DefaultActorClock in all threads this might cause issues on updating the time https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/clock/DefaultActorClock.java#L24-L31 which is not thread safe and called by all threads https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L78 and also when timers are scheduled https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorTimerQueue.java#L53 I found that during investigating https://github.com/camunda/zeebe/pull/10390 and trying to make ActorTimerQueue threadsafe... <!-- A clear and concise description of what the bug is. --> **To Reproduce** <!-- Steps to reproduce the behavior If possible add a minimal reproducer code sample - when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java --> **Expected behavior** Either we not set the clock or we use a thread safe structure. <!-- A clear and concise description of what you expected to happen. --> **Log/Stacktrace** <!-- If possible add the full stacktrace or Zeebe log which contains the issue. --> <details><summary>Full Stacktrace</summary> <p> ``` <STACKTRACE> ``` </p> </details> **Environment:** - OS: <!-- [e.g. Linux] --> - Zeebe Version: 8.x<!-- [e.g. 0.20.0] --> - Configuration: <!-- [e.g. exporters etc.] -->
True
DefaultActorClock is not thread safe - **Describe the bug** It seems we now create the ActorScheduler with a predefined actor clock, either ControlledClock or the DefaultActorClock https://github.com/camunda/zeebe/blob/main/dist/src/main/java/io/camunda/zeebe/shared/ActorClockConfiguration.java#L34-L35 The problem is that the DefaultActorClock is not thread safe! Previously we created the ActorScheduler without clock, which caused [that each thread has its own clock](https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L64). Then there is no issue. If we use the DefaultActorClock in all threads this might cause issues on updating the time https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/clock/DefaultActorClock.java#L24-L31 which is not thread safe and called by all threads https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L78 and also when timers are scheduled https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorTimerQueue.java#L53 I found that during investigating https://github.com/camunda/zeebe/pull/10390 and trying to make ActorTimerQueue threadsafe... <!-- A clear and concise description of what the bug is. --> **To Reproduce** <!-- Steps to reproduce the behavior If possible add a minimal reproducer code sample - when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java --> **Expected behavior** Either we not set the clock or we use a thread safe structure. <!-- A clear and concise description of what you expected to happen. --> **Log/Stacktrace** <!-- If possible add the full stacktrace or Zeebe log which contains the issue. --> <details><summary>Full Stacktrace</summary> <p> ``` <STACKTRACE> ``` </p> </details> **Environment:** - OS: <!-- [e.g. Linux] --> - Zeebe Version: 8.x<!-- [e.g. 0.20.0] --> - Configuration: <!-- [e.g. exporters etc.] -->
reli
defaultactorclock is not thread safe describe the bug it seems we now create the actorscheduler with a predefined actor clock either controlledclock or the defaultactorclock the problem is that the defaultactorclock is not thread safe previously we created the actorscheduler without clock which caused then there is no issue if we use the defaultactorclock in all threads this might cause issues on updating the time which is not thread safe and called by all threads and also when timers are scheduled i found that during investigating and trying to make actortimerqueue threadsafe to reproduce steps to reproduce the behavior if possible add a minimal reproducer code sample when using the java client expected behavior either we not set the clock or we use a thread safe structure log stacktrace full stacktrace environment os zeebe version x configuration
1
687,162
23,515,782,898
IssuesEvent
2022-08-18 21:11:51
ezolenko/rollup-plugin-typescript2
https://api.github.com/repos/ezolenko/rollup-plugin-typescript2
closed
some type-only TS files are ignored when using `tsconfig` `files`?
kind: bug solution: duplicate problem: removed issue template priority: in progress problem: plugin order topic: type-only / emit-less imports
## What happens and why it is incorrect there are many files ending in `*.ts` in the src, but only some have been transpiled 1. Does `tsc` have the same output? If so, please explain why this is incorrect behavior NO, `tsc` run as expectly. ![image](https://user-images.githubusercontent.com/52886395/185072389-587c5bfa-0817-4c06-a4c9-3a5833ba1457.png) here's file structure: ![image](https://user-images.githubusercontent.com/52886395/185073553-3a473e39-d5ae-41ea-b5dc-f32a5e10a4eb.png) but rpt2 seems to ignore the file named `component.ts`: <img width="1241" alt="image" src="https://user-images.githubusercontent.com/52886395/185072856-019e3e3e-9598-4a5e-8dfb-10cb9f7276cc.png"> ## Environment ### Versions ```text System: OS: macOS 11.4 CPU: (12) x64 Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz Memory: 59.11 MB / 32.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 14.17.4 - ~/.nvm/versions/node/v14.17.4/bin/node Yarn: 1.22.11 - ~/.nvm/versions/node/v14.17.4/bin/yarn npm: 6.14.14 - ~/.nvm/versions/node/v14.17.4/bin/npm npmPackages: rollup: ^2.45.2 => 2.45.2 rollup-plugin-typescript2: ^0.30.0 => 0.30.0 typescript: ^4.3.5 => 4.3.5 npmGlobalPackages: typescript: 4.3.5 ``` <!--- paste your rollup config below if relevant ---> <details> <summary><h4><code>rollup.config.js</code></h4>: </summary> <!--- INSERT rollup.config.ts IN THE CODE SNIPPET BELOW ---> ```js import { nodeResolve } from '@rollup/plugin-node-resolve'; import replace from '@rollup/plugin-replace'; import postcss from 'rollup-plugin-postcss'; import image from '@rollup/plugin-image'; import ts from 'rollup-plugin-typescript2'; import dts from 'rollup-plugin-dts'; import json from '@rollup/plugin-json'; import alias from '@rollup/plugin-alias'; import autoExternal from 'rollup-plugin-auto-external'; import path from 'path'; import importCss from './rollup-plugin-import-css'; const resolve = (...dirs) => path.resolve(__dirname, '../', ...dirs); function toUpperCase(match) { return match.replace('-', '').toUpperCase(); } export default [ { input: resolve('src/lib/index.tsx'), output: [ { file: resolve('dist/index.esm.js'), format: 'es', sourcemap: true, }, ], external: id => { try { const idSourcePath = require.resolve(id, { paths: [resolve()] }); return idSourcePath && idSourcePath.includes('node_modules'); } catch (error) { return false; } }, plugins: [ ts({ check: false, tsconfig: resolve('tsconfig.json'), tsconfigOverride: { compilerOptions: { declaration: true, declarationDir: resolve('dist/type') }, }, verbosity: 2, useTsconfigDeclarationDir: true, include: ['*.ts+(|x)', '**/*.ts+(|x)', '*.js+(|x)', '**/*.js+(|x)'], }), image(), postcss({ extensions: ['.css', '.scss', '.less'], autoModules: true, extract: 'index.css', namedExports(name) { let reg = /-[a-z]/g; const temp = name.replace(reg, toUpperCase); return temp; }, }), json(), alias({ entries: [ { find: 'src', replacement: resolve('src'), }, ], }), nodeResolve({ extensions: ['.js', '.jsx', '.ts', '.tsx'], // some package.json files have a "browser" field which specifies // alternative files to load for people bundling for the browser. If // that's you, either use this option or add "browser" to the // "mainfields" option, otherwise pkg.browser will be ignored browser: true, preferBuiltins: true, mainFields: ['browser', 'jsnext', 'module', 'main'], }), replace({ 'process.env.NODE_ENV': JSON.stringify('development'), 'process.env.SEMI_ICON_LAZY_LOAD': true, preventAssignment: true, }), autoExternal({ packagePath: resolve(), }), importCss(), ], }, { input: resolve('dist/type/src/lib/index.d.ts'), external: id => { try { const idSourcePath = require.resolve(id, { paths: [resolve()] }); return idSourcePath && idSourcePath.includes('node_modules'); } catch (error) { return false; } }, output: [{ file: resolve('dist/index.d.ts'), format: 'es' }], plugins: [dts()], }, ]; ``` </details> <!--- paste your tsconfig.json below if relevant ---> <details> <summary><h4><code>tsconfig.json</code></h4>: </summary> <!--- INSERT tsconfig.json IN THE CODE SNIPPET BELOW ---> ```json5 { "compilerOptions": { "target": "es6", "module": "esnext", "lib": ["es7", "dom"], "sourceMap": true, "allowJs": true, "jsx": "react", "moduleResolution": "node", "experimentalDecorators": true, "rootDir": "./", "baseUrl": "./src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "noEmit": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-editor": ["src/lib/*"] }, "plugins": [ { "transform": "ts-optchain/transform" } ] }, "typeRoots": ["node", "node_modules/@types", "./src/typings"], "exclude": ["node_modules"] } ``` </details> <!--- paste your package.json below if relevant ---> <details> <summary><h4><code>package.json</code></h4>: </summary> <!--- INSERT package.json IN THE CODE SNIPPET BELOW ---> ```json ``` </details> <!--- add verbosity verbosity: 3 to plugin options and attach output if relevant (censor out anything sensitive) ---> <details> <summary><h4>plugin output with verbosity 3</h4>: </summary> <!--- INSERT plugin output IN THE CODE SNIPPET BELOW or attach ---> ```text /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx → dist/index.esm.js... rpt2: built-in options overrides: { "noEmitHelpers": false, "importHelpers": true, "noResolve": false, "noEmit": false, "inlineSourceMap": false, "outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder", "moduleResolution": 2, "allowNonTsExtensions": true } rpt2: parsed tsconfig: { "options": { "isolatedModules": false, "declaration": true, "target": 2, "module": 99, "lib": [ "lib.es2016.d.ts", "lib.dom.d.ts" ], "sourceMap": true, "allowJs": true, "jsx": 2, "moduleResolution": 2, "experimentalDecorators": true, "rootDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "baseUrl": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-tag-table": [ "src/lib/*" ] }, "plugins": [ { "transform": "ts-optchain/transform" } ], "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type", "configFilePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json", "pathsBasePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "noEmitHelpers": false, "noResolve": false, "noEmit": false, "inlineSourceMap": false, "outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder", "allowNonTsExtensions": true }, "fileNames": [ "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx" ], "typeAcquisition": { "enable": false, "include": [], "exclude": [] }, "raw": { "files": [ "src/lib/index.tsx" ], "compilerOptions": { "isolatedModules": false, "declaration": true, "target": "es6", "module": "esnext", "lib": [ "es7", "dom" ], "sourceMap": true, "allowJs": true, "jsx": "react", "moduleResolution": "node", "experimentalDecorators": true, "rootDir": "./", "baseUrl": "./src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-tag-table": [ "src/lib/*" ] }, "plugins": [ { "transform": "ts-optchain/transform" } ], "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type" }, "typeRoots": [ "node", "node_modules/@types", "./src/typings" ], "exclude": [ "node_modules" ], "compileOnSave": false }, "errors": [], "wildcardDirectories": {}, "compileOnSave": false } rpt2: typescript version: 4.3.5 rpt2: tslib version: 2.1.0 rpt2: rollup version: 2.45.2 rpt2: rollup-plugin-typescript2 version: 0.30.0 rpt2: plugin options: { "check": false, "tsconfig": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json", "tsconfigOverride": { "compilerOptions": { "declaration": true, "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type" } }, "verbosity": 3, "useTsconfigDeclarationDir": true, "include": [ "*.ts+(|x)", "**/*.ts+(|x)", "*.js+(|x)", "**/*.js+(|x)" ], "clean": false, "cacheRoot": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2", "exclude": [ "*.d.ts", "**/*.d.ts" ], "abortOnError": true, "rollupCommonJSResolveHack": false, "transformers": [], "tsconfigDefaults": {}, "objectHashIgnoreUnknownHack": false, "cwd": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "typescript": "version 4.3.5" } rpt2: rollup config: { "input": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx", "plugins": [ { "name": "rpt2" }, { "name": "image" }, { "name": "postcss" }, { "name": "json" }, { "name": "alias" }, { "name": "node-resolve" }, { "name": "replace" }, { "name": "auto-external" }, {}, { "name": "stdin" } ], "output": [ { "file": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/index.esm.js", "format": "es", "plugins": [], "sourcemap": true } ] } rpt2: tsconfig path: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json rpt2: included: [ "*.ts+(|x)", "**/*.ts+(|x)", "*.js+(|x)", "**/*.js+(|x)" ] rpt2: excluded: [ "*.d.ts", "**/*.d.ts" ] rpt2: Ambient types: rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__core/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__generator/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__template/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__traverse/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/color-name/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/css-modules-loader-core/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/estree/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/graceful-fs/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/hast/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-coverage/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-report/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-reports/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/jest/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/mdast/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse5/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prettier/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prop-types/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/q/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react-dom/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/resolve/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/stack-utils/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/unist/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs-parser/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts rpt2: /Users/bytedance/node_modules/@types/json5/index.d.ts rpt2: /Users/bytedance/node_modules/@types/minimist/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts rpt2: ambient types changed, redoing all semantic diagnostics rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/7cdbf1c7db69c2fd53815824bca49513a129c642' rpt2: cache miss rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: resolving './utils' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: resolving './mock' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/a3be5c51ebb3fef03fc9a43c539d188ad1577216' rpt2: cache hit rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/d1d6e9acec5e9e0b4392fcbc5bf623dab081aa31' rpt2: cache hit rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: generating target 1 rpt2: rolling caches rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/index.d.ts' rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/mock.d.ts' rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/utils.d.ts' (!) Broken sourcemap https://rollupjs.org/guide/en/#warning-sourcemap-is-likely-to-be-incorrect Plugins that transform code (such as 'at position 9') should generate accompanying sourcemaps created dist/index.esm.js in 2.7s User/dist/type/src/lib/index.d.ts → dist/index.d.ts... [!] Error: Could not resolve './component' from dist/type/src/lib/index.d.ts Error: Could not resolve './component' from dist/type/src/lib/index.d.ts ``` </details>
1.0
some type-only TS files are ignored when using `tsconfig` `files`? - ## What happens and why it is incorrect there are many files ending in `*.ts` in the src, but only some have been transpiled 1. Does `tsc` have the same output? If so, please explain why this is incorrect behavior NO, `tsc` run as expectly. ![image](https://user-images.githubusercontent.com/52886395/185072389-587c5bfa-0817-4c06-a4c9-3a5833ba1457.png) here's file structure: ![image](https://user-images.githubusercontent.com/52886395/185073553-3a473e39-d5ae-41ea-b5dc-f32a5e10a4eb.png) but rpt2 seems to ignore the file named `component.ts`: <img width="1241" alt="image" src="https://user-images.githubusercontent.com/52886395/185072856-019e3e3e-9598-4a5e-8dfb-10cb9f7276cc.png"> ## Environment ### Versions ```text System: OS: macOS 11.4 CPU: (12) x64 Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz Memory: 59.11 MB / 32.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 14.17.4 - ~/.nvm/versions/node/v14.17.4/bin/node Yarn: 1.22.11 - ~/.nvm/versions/node/v14.17.4/bin/yarn npm: 6.14.14 - ~/.nvm/versions/node/v14.17.4/bin/npm npmPackages: rollup: ^2.45.2 => 2.45.2 rollup-plugin-typescript2: ^0.30.0 => 0.30.0 typescript: ^4.3.5 => 4.3.5 npmGlobalPackages: typescript: 4.3.5 ``` <!--- paste your rollup config below if relevant ---> <details> <summary><h4><code>rollup.config.js</code></h4>: </summary> <!--- INSERT rollup.config.ts IN THE CODE SNIPPET BELOW ---> ```js import { nodeResolve } from '@rollup/plugin-node-resolve'; import replace from '@rollup/plugin-replace'; import postcss from 'rollup-plugin-postcss'; import image from '@rollup/plugin-image'; import ts from 'rollup-plugin-typescript2'; import dts from 'rollup-plugin-dts'; import json from '@rollup/plugin-json'; import alias from '@rollup/plugin-alias'; import autoExternal from 'rollup-plugin-auto-external'; import path from 'path'; import importCss from './rollup-plugin-import-css'; const resolve = (...dirs) => path.resolve(__dirname, '../', ...dirs); function toUpperCase(match) { return match.replace('-', '').toUpperCase(); } export default [ { input: resolve('src/lib/index.tsx'), output: [ { file: resolve('dist/index.esm.js'), format: 'es', sourcemap: true, }, ], external: id => { try { const idSourcePath = require.resolve(id, { paths: [resolve()] }); return idSourcePath && idSourcePath.includes('node_modules'); } catch (error) { return false; } }, plugins: [ ts({ check: false, tsconfig: resolve('tsconfig.json'), tsconfigOverride: { compilerOptions: { declaration: true, declarationDir: resolve('dist/type') }, }, verbosity: 2, useTsconfigDeclarationDir: true, include: ['*.ts+(|x)', '**/*.ts+(|x)', '*.js+(|x)', '**/*.js+(|x)'], }), image(), postcss({ extensions: ['.css', '.scss', '.less'], autoModules: true, extract: 'index.css', namedExports(name) { let reg = /-[a-z]/g; const temp = name.replace(reg, toUpperCase); return temp; }, }), json(), alias({ entries: [ { find: 'src', replacement: resolve('src'), }, ], }), nodeResolve({ extensions: ['.js', '.jsx', '.ts', '.tsx'], // some package.json files have a "browser" field which specifies // alternative files to load for people bundling for the browser. If // that's you, either use this option or add "browser" to the // "mainfields" option, otherwise pkg.browser will be ignored browser: true, preferBuiltins: true, mainFields: ['browser', 'jsnext', 'module', 'main'], }), replace({ 'process.env.NODE_ENV': JSON.stringify('development'), 'process.env.SEMI_ICON_LAZY_LOAD': true, preventAssignment: true, }), autoExternal({ packagePath: resolve(), }), importCss(), ], }, { input: resolve('dist/type/src/lib/index.d.ts'), external: id => { try { const idSourcePath = require.resolve(id, { paths: [resolve()] }); return idSourcePath && idSourcePath.includes('node_modules'); } catch (error) { return false; } }, output: [{ file: resolve('dist/index.d.ts'), format: 'es' }], plugins: [dts()], }, ]; ``` </details> <!--- paste your tsconfig.json below if relevant ---> <details> <summary><h4><code>tsconfig.json</code></h4>: </summary> <!--- INSERT tsconfig.json IN THE CODE SNIPPET BELOW ---> ```json5 { "compilerOptions": { "target": "es6", "module": "esnext", "lib": ["es7", "dom"], "sourceMap": true, "allowJs": true, "jsx": "react", "moduleResolution": "node", "experimentalDecorators": true, "rootDir": "./", "baseUrl": "./src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "noEmit": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-editor": ["src/lib/*"] }, "plugins": [ { "transform": "ts-optchain/transform" } ] }, "typeRoots": ["node", "node_modules/@types", "./src/typings"], "exclude": ["node_modules"] } ``` </details> <!--- paste your package.json below if relevant ---> <details> <summary><h4><code>package.json</code></h4>: </summary> <!--- INSERT package.json IN THE CODE SNIPPET BELOW ---> ```json ``` </details> <!--- add verbosity verbosity: 3 to plugin options and attach output if relevant (censor out anything sensitive) ---> <details> <summary><h4>plugin output with verbosity 3</h4>: </summary> <!--- INSERT plugin output IN THE CODE SNIPPET BELOW or attach ---> ```text /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx → dist/index.esm.js... rpt2: built-in options overrides: { "noEmitHelpers": false, "importHelpers": true, "noResolve": false, "noEmit": false, "inlineSourceMap": false, "outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder", "moduleResolution": 2, "allowNonTsExtensions": true } rpt2: parsed tsconfig: { "options": { "isolatedModules": false, "declaration": true, "target": 2, "module": 99, "lib": [ "lib.es2016.d.ts", "lib.dom.d.ts" ], "sourceMap": true, "allowJs": true, "jsx": 2, "moduleResolution": 2, "experimentalDecorators": true, "rootDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "baseUrl": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-tag-table": [ "src/lib/*" ] }, "plugins": [ { "transform": "ts-optchain/transform" } ], "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type", "configFilePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json", "pathsBasePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "noEmitHelpers": false, "noResolve": false, "noEmit": false, "inlineSourceMap": false, "outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder", "allowNonTsExtensions": true }, "fileNames": [ "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx" ], "typeAcquisition": { "enable": false, "include": [], "exclude": [] }, "raw": { "files": [ "src/lib/index.tsx" ], "compilerOptions": { "isolatedModules": false, "declaration": true, "target": "es6", "module": "esnext", "lib": [ "es7", "dom" ], "sourceMap": true, "allowJs": true, "jsx": "react", "moduleResolution": "node", "experimentalDecorators": true, "rootDir": "./", "baseUrl": "./src", "forceConsistentCasingInFileNames": true, "noImplicitReturns": true, "noImplicitThis": false, "noImplicitAny": false, "importHelpers": true, "strictNullChecks": false, "suppressImplicitAnyIndexErrors": true, "noUnusedLocals": true, "allowSyntheticDefaultImports": true, "esModuleInterop": false, "paths": { "@ies/kefu-tag-table": [ "src/lib/*" ] }, "plugins": [ { "transform": "ts-optchain/transform" } ], "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type" }, "typeRoots": [ "node", "node_modules/@types", "./src/typings" ], "exclude": [ "node_modules" ], "compileOnSave": false }, "errors": [], "wildcardDirectories": {}, "compileOnSave": false } rpt2: typescript version: 4.3.5 rpt2: tslib version: 2.1.0 rpt2: rollup version: 2.45.2 rpt2: rollup-plugin-typescript2 version: 0.30.0 rpt2: plugin options: { "check": false, "tsconfig": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json", "tsconfigOverride": { "compilerOptions": { "declaration": true, "declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type" } }, "verbosity": 3, "useTsconfigDeclarationDir": true, "include": [ "*.ts+(|x)", "**/*.ts+(|x)", "*.js+(|x)", "**/*.js+(|x)" ], "clean": false, "cacheRoot": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2", "exclude": [ "*.d.ts", "**/*.d.ts" ], "abortOnError": true, "rollupCommonJSResolveHack": false, "transformers": [], "tsconfigDefaults": {}, "objectHashIgnoreUnknownHack": false, "cwd": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable", "typescript": "version 4.3.5" } rpt2: rollup config: { "input": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx", "plugins": [ { "name": "rpt2" }, { "name": "image" }, { "name": "postcss" }, { "name": "json" }, { "name": "alias" }, { "name": "node-resolve" }, { "name": "replace" }, { "name": "auto-external" }, {}, { "name": "stdin" } ], "output": [ { "file": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/index.esm.js", "format": "es", "plugins": [], "sourcemap": true } ] } rpt2: tsconfig path: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json rpt2: included: [ "*.ts+(|x)", "**/*.ts+(|x)", "*.js+(|x)", "**/*.js+(|x)" ] rpt2: excluded: [ "*.d.ts", "**/*.d.ts" ] rpt2: Ambient types: rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__core/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__generator/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__template/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__traverse/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/color-name/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/css-modules-loader-core/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/estree/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/graceful-fs/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/hast/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-coverage/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-report/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-reports/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/jest/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/mdast/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse5/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prettier/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prop-types/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/q/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react-dom/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/resolve/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/stack-utils/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/unist/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs-parser/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts rpt2: /Users/bytedance/node_modules/@types/json5/index.d.ts rpt2: /Users/bytedance/node_modules/@types/minimist/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts rpt2: ambient types changed, redoing all semantic diagnostics rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/7cdbf1c7db69c2fd53815824bca49513a129c642' rpt2: cache miss rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: resolving './utils' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: resolving './mock' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/a3be5c51ebb3fef03fc9a43c539d188ad1577216' rpt2: cache hit rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/d1d6e9acec5e9e0b4392fcbc5bf623dab081aa31' rpt2: cache hit rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' rpt2: generating target 1 rpt2: rolling caches rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/index.d.ts' rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/mock.d.ts' rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/utils.d.ts' (!) Broken sourcemap https://rollupjs.org/guide/en/#warning-sourcemap-is-likely-to-be-incorrect Plugins that transform code (such as 'at position 9') should generate accompanying sourcemaps created dist/index.esm.js in 2.7s User/dist/type/src/lib/index.d.ts → dist/index.d.ts... [!] Error: Could not resolve './component' from dist/type/src/lib/index.d.ts Error: Could not resolve './component' from dist/type/src/lib/index.d.ts ``` </details>
non_reli
some type only ts files are ignored when using tsconfig files what happens and why it is incorrect there are many files ending in ts in the src but only some have been transpiled does tsc have the same output if so please explain why this is incorrect behavior no, tsc run as expectly here s file structure but seems to ignore the file named component ts img width alt image src environment versions text system os macos cpu intel r core tm cpu memory mb gb shell bin zsh binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm npmpackages rollup rollup plugin typescript npmglobalpackages typescript rollup config js js import noderesolve from rollup plugin node resolve import replace from rollup plugin replace import postcss from rollup plugin postcss import image from rollup plugin image import ts from rollup plugin import dts from rollup plugin dts import json from rollup plugin json import alias from rollup plugin alias import autoexternal from rollup plugin auto external import path from path import importcss from rollup plugin import css const resolve dirs path resolve dirname dirs function touppercase match return match replace touppercase export default input resolve src lib index tsx output file resolve dist index esm js format es sourcemap true external id try const idsourcepath require resolve id paths return idsourcepath idsourcepath includes node modules catch error return false plugins ts check false tsconfig resolve tsconfig json tsconfigoverride compileroptions declaration true declarationdir resolve dist type verbosity usetsconfigdeclarationdir true include image postcss extensions automodules true extract index css namedexports name let reg g const temp name replace reg touppercase return temp json alias entries find src replacement resolve src noderesolve extensions some package json files have a browser field which specifies alternative files to load for people bundling for the browser if that s you either use this option or add browser to the mainfields option otherwise pkg browser will be ignored browser true preferbuiltins true mainfields replace process env node env json stringify development process env semi icon lazy load true preventassignment true autoexternal packagepath resolve importcss input resolve dist type src lib index d ts external id try const idsourcepath require resolve id paths return idsourcepath idsourcepath includes node modules catch error return false output plugins tsconfig json compileroptions target module esnext lib sourcemap true allowjs true jsx react moduleresolution node experimentaldecorators true rootdir baseurl src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true noemit true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu editor plugins transform ts optchain transform typeroots exclude package json json plugin output with verbosity text users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx → dist index esm js built in options overrides noemithelpers false importhelpers true noresolve false noemit false inlinesourcemap false outdir users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin placeholder moduleresolution allownontsextensions true parsed tsconfig options isolatedmodules false declaration true target module lib lib d ts lib dom d ts sourcemap true allowjs true jsx moduleresolution experimentaldecorators true rootdir users bytedance public helpdesk semi ui packages kefucascadertable baseurl users bytedance public helpdesk semi ui packages kefucascadertable src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu tag table src lib plugins transform ts optchain transform declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type configfilepath users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json pathsbasepath users bytedance public helpdesk semi ui packages kefucascadertable noemithelpers false noresolve false noemit false inlinesourcemap false outdir users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin placeholder allownontsextensions true filenames users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx typeacquisition enable false include exclude raw files src lib index tsx compileroptions isolatedmodules false declaration true target module esnext lib dom sourcemap true allowjs true jsx react moduleresolution node experimentaldecorators true rootdir baseurl src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu tag table src lib plugins transform ts optchain transform declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type typeroots node node modules types src typings exclude node modules compileonsave false errors wildcarddirectories compileonsave false typescript version tslib version rollup version rollup plugin version plugin options check false tsconfig users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json tsconfigoverride compileroptions declaration true declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type verbosity usetsconfigdeclarationdir true include ts x ts x js x js x clean false cacheroot users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin exclude d ts d ts abortonerror true rollupcommonjsresolvehack false transformers tsconfigdefaults objecthashignoreunknownhack false cwd users bytedance public helpdesk semi ui packages kefucascadertable typescript version rollup config input users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx plugins name name image name postcss name json name alias name node resolve name replace name auto external name stdin output file users bytedance public helpdesk semi ui packages kefucascadertable dist index esm js format es plugins sourcemap true tsconfig path users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json included ts x ts x js x js x excluded d ts d ts ambient types users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel core index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel generator index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel template index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel traverse index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types color name index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types css modules loader core index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types estree index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types graceful fs index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types hast index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul lib coverage index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul lib report index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul reports index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types jest index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types json schema index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types mdast index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types node index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types normalize package data index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types parse json index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types prettier index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types prop types index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types q index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types react index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types react dom index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types resolve index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types stack utils index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types unist index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types yargs index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types yargs parser index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types json schema index d ts users bytedance node modules types index d ts users bytedance node modules types minimist index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types node index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types normalize package data index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types parse json index d ts ambient types changed redoing all semantic diagnostics transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache miss generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx dependency users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx resolving utils imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts dependency users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx resolving mock imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache hit generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache hit generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts generating target rolling caches emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib index d ts emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib mock d ts emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib utils d ts broken sourcemap plugins that transform code such as at position should generate accompanying sourcemaps created dist index esm js in user dist type src lib index d ts → dist index d ts error could not resolve component from dist type src lib index d ts error could not resolve component from dist type src lib index d ts
0
2,862
28,365,904,500
IssuesEvent
2023-04-12 13:53:15
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
closed
[ICS02] `ClientStatePath` is not checked during client creation
O: reliability I: specs
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statment - The [ICS02](https://github.com/cosmos/ibc/tree/main/spec/core/ics-002-client-semantics#create) specifies that for handling a client creation request, the process should check if the submitted `ClientState` already exists in the `ClientStatePath`! > abortTransactionUnless(provableStore.get(clientStatePath(identifier)) === null) - Note that an incorrect use of `SetClientState` can result in a `ClientState` being set to the same store as the one that is going to be set by a `MsgCreateClient`. ## In IBC-go - There is a [`GetClientState`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/keeper.go#L79-L89) method, but **not** used to perform this check. Instead they create an isolated store by [`StoreClient`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/client.go#L32) for this purpose. - This can't be applied to IBC-rs, as we do not access to the store. ## Acceptance Criteria Check `ClientStatePath` to ensure the received `ClientState` via a `MsgCreateClient` does not exist
True
[ICS02] `ClientStatePath` is not checked during client creation - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statment - The [ICS02](https://github.com/cosmos/ibc/tree/main/spec/core/ics-002-client-semantics#create) specifies that for handling a client creation request, the process should check if the submitted `ClientState` already exists in the `ClientStatePath`! > abortTransactionUnless(provableStore.get(clientStatePath(identifier)) === null) - Note that an incorrect use of `SetClientState` can result in a `ClientState` being set to the same store as the one that is going to be set by a `MsgCreateClient`. ## In IBC-go - There is a [`GetClientState`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/keeper.go#L79-L89) method, but **not** used to perform this check. Instead they create an isolated store by [`StoreClient`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/client.go#L32) for this purpose. - This can't be applied to IBC-rs, as we do not access to the store. ## Acceptance Criteria Check `ClientStatePath` to ensure the received `ClientState` via a `MsgCreateClient` does not exist
reli
clientstatepath is not checked during client creation ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ problem statment the specifies that for handling a client creation request the process should check if the submitted clientstate already exists in the clientstatepath aborttransactionunless provablestore get clientstatepath identifier null note that an incorrect use of setclientstate can result in a clientstate being set to the same store as the one that is going to be set by a msgcreateclient in ibc go there is a method but not used to perform this check instead they create an isolated store by for this purpose this can t be applied to ibc rs as we do not access to the store acceptance criteria check clientstatepath to ensure the received clientstate via a msgcreateclient does not exist
1
20,356
13,879,441,325
IssuesEvent
2020-10-17 14:28:20
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
closed
Error when starting local dev env
Pod: WP & Infra Status: Not Applicable Type: Bug Type: Infrastructure
## Bug Description When starting local dev environment I get the following error message. ``` STATUS: Import default set of media assets... Imported file '/var/www/html/wp-content/e2e-assets/example-1.jpg' as attachment ID 1965. Success: Imported 1 of 1 items. Imported file '/var/www/html/wp-content/e2e-assets/example-2.jpg' as attachment ID 1966. Success: Imported 1 of 1 items. Imported file '/var/www/html/wp-content/e2e-assets/example-3.png' as attachment ID 1967. Success: Imported 1 of 1 items. STATUS: Current WordPress version: 5.6-alpha-49085... STATUS: Updating WordPress to the latest version... [05-Oct-2020 11:25:06 UTC] PHP Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30 Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30 ``` ## Expected Behaviour It works. ## Steps to Reproduce Type `npm run env:start` ## Screenshots <!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! --> ## Additional Context <!-- Please complete the following information. --> - Plugin Version: 1-RC1-1.0.0 - Operating System: Mac OS 10.15.7 --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
1.0
Error when starting local dev env - ## Bug Description When starting local dev environment I get the following error message. ``` STATUS: Import default set of media assets... Imported file '/var/www/html/wp-content/e2e-assets/example-1.jpg' as attachment ID 1965. Success: Imported 1 of 1 items. Imported file '/var/www/html/wp-content/e2e-assets/example-2.jpg' as attachment ID 1966. Success: Imported 1 of 1 items. Imported file '/var/www/html/wp-content/e2e-assets/example-3.png' as attachment ID 1967. Success: Imported 1 of 1 items. STATUS: Current WordPress version: 5.6-alpha-49085... STATUS: Updating WordPress to the latest version... [05-Oct-2020 11:25:06 UTC] PHP Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30 Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30 ``` ## Expected Behaviour It works. ## Steps to Reproduce Type `npm run env:start` ## Screenshots <!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! --> ## Additional Context <!-- Please complete the following information. --> - Plugin Version: 1-RC1-1.0.0 - Operating System: Mac OS 10.15.7 --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
non_reli
error when starting local dev env bug description when starting local dev environment i get the following error message status import default set of media assets imported file var www html wp content assets example jpg as attachment id success imported of items imported file var www html wp content assets example jpg as attachment id success imported of items imported file var www html wp content assets example png as attachment id success imported of items status current wordpress version alpha status updating wordpress to the latest version php warning declaration of wp cli core coreupgrader download package package check signatures true should be compatible with wp upgrader download package package check signatures false hook extra array in phar usr local bin wp vendor wp cli core command src wp cli core coreupgrader php on line warning declaration of wp cli core coreupgrader download package package check signatures true should be compatible with wp upgrader download package package check signatures false hook extra array in phar usr local bin wp vendor wp cli core command src wp cli core coreupgrader php on line expected behaviour it works steps to reproduce type npm run env start screenshots additional context plugin version operating system mac os do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief
0
392,678
11,594,442,126
IssuesEvent
2020-02-24 15:20:31
arfc/2020-dotson-optimal-sizing
https://api.github.com/repos/arfc/2020-dotson-optimal-sizing
opened
Convert steam flow rates to thermal energy requirement.
Difficulty:1-Beginner Priority:2-Normal Status:4-In Progress Type:Feature
This issue can be closed when - [ ] mass flow rate of steam has been converted to thermal energy requirements - [ ] sythetic histories have been generated
1.0
Convert steam flow rates to thermal energy requirement. - This issue can be closed when - [ ] mass flow rate of steam has been converted to thermal energy requirements - [ ] sythetic histories have been generated
non_reli
convert steam flow rates to thermal energy requirement this issue can be closed when mass flow rate of steam has been converted to thermal energy requirements sythetic histories have been generated
0
272
5,938,939,826
IssuesEvent
2017-05-25 02:27:54
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
VS Crash: contract failed in CodeModel.AbstractProjectCodeModel.GetCodeModelCache
Area-IDE Bug Tenet-Reliability
While working in Compilers.sln. Build: D15Rel 26202.1 Stack trace: ``` at Roslyn.Utilities.Contract.ThrowIfFalse(Boolean condition, String message) in E:\A\_work\6\s\src\Workspaces\Core\Portable\Utilities\Contract.cs:line 86 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.GetCodeModelCache() in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 54 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.TryGetCachedFileCodeModel(String fileName, ComHandle`2& fileCodeModelHandle) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 69 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.FireEvents(DocumentId documentId, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 86 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.AnalyzeSyntaxAsync(Document document, InvocationReasons reasons, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 60 at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass30_0.<ProcessDocumentAnalyzersAsync>b__0(IIncrementalAnalyzer a, Document d, CancellationToken c) in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 153 at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass31_1`1.<<RunAnalyzersAsync>b__0>d.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 187 at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<GetOrDefaultAsync>d__33`2.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 229 ``` Memory dump: ```%internal_share%\public\tomat\Bugs\16973```
True
VS Crash: contract failed in CodeModel.AbstractProjectCodeModel.GetCodeModelCache - While working in Compilers.sln. Build: D15Rel 26202.1 Stack trace: ``` at Roslyn.Utilities.Contract.ThrowIfFalse(Boolean condition, String message) in E:\A\_work\6\s\src\Workspaces\Core\Portable\Utilities\Contract.cs:line 86 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.GetCodeModelCache() in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 54 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.TryGetCachedFileCodeModel(String fileName, ComHandle`2& fileCodeModelHandle) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 69 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.FireEvents(DocumentId documentId, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 86 at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.AnalyzeSyntaxAsync(Document document, InvocationReasons reasons, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 60 at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass30_0.<ProcessDocumentAnalyzersAsync>b__0(IIncrementalAnalyzer a, Document d, CancellationToken c) in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 153 at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass31_1`1.<<RunAnalyzersAsync>b__0>d.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 187 at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<GetOrDefaultAsync>d__33`2.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 229 ``` Memory dump: ```%internal_share%\public\tomat\Bugs\16973```
reli
vs crash contract failed in codemodel abstractprojectcodemodel getcodemodelcache while working in compilers sln build stack trace at roslyn utilities contract throwiffalse boolean condition string message in e a work s src workspaces core portable utilities contract cs line at microsoft visualstudio languageservices implementation codemodel abstractprojectcodemodel getcodemodelcache in e a work s src visualstudio core impl codemodel abstractprojectcodemodel cs line at microsoft visualstudio languageservices implementation codemodel abstractprojectcodemodel trygetcachedfilecodemodel string filename comhandle filecodemodelhandle in e a work s src visualstudio core impl codemodel abstractprojectcodemodel cs line at microsoft visualstudio languageservices implementation codemodel codemodelincrementalanalyzerprovider analyzer fireevents documentid documentid cancellationtoken cancellationtoken in e a work s src visualstudio core impl codemodel codemodelincrementalanalyzer cs line at microsoft visualstudio languageservices implementation codemodel codemodelincrementalanalyzerprovider analyzer analyzesyntaxasync document document invocationreasons reasons cancellationtoken cancellationtoken in e a work s src visualstudio core impl codemodel codemodelincrementalanalyzer cs line at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor c b iincrementalanalyzer a document d cancellationtoken c in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor c b d movenext in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor d movenext in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line memory dump internal share public tomat bugs
1
1,374
15,606,232,799
IssuesEvent
2021-03-19 07:42:05
FoundationDB/fdb-kubernetes-operator
https://api.github.com/repos/FoundationDB/fdb-kubernetes-operator
closed
Logging all decisions to remove processes
reliability
We've seen some cases where processes were marked for removal and it wasn't clear why. We should add a log statement whenever we mark a process for removal identifying the reason, so we can search for this event in the logs when we see unexplained removals.
True
Logging all decisions to remove processes - We've seen some cases where processes were marked for removal and it wasn't clear why. We should add a log statement whenever we mark a process for removal identifying the reason, so we can search for this event in the logs when we see unexplained removals.
reli
logging all decisions to remove processes we ve seen some cases where processes were marked for removal and it wasn t clear why we should add a log statement whenever we mark a process for removal identifying the reason so we can search for this event in the logs when we see unexplained removals
1
2,743
27,378,169,223
IssuesEvent
2023-02-28 08:05:41
jasp-stats/jasp-issues
https://api.github.com/repos/jasp-stats/jasp-issues
closed
Intercalas correlation coefficient
Module: jaspReliability
* Enhancement: add the ICC calculation in the reliability analysis section * Purpose: to calculate the ICC for reliability studies with multiple quantitative variables * Use-case: in sports science is typically asses the reliability of instruments that report quantitative variables, nowadays we need to compute this calculation using other statistical packages like SPSS.
True
Intercalas correlation coefficient - * Enhancement: add the ICC calculation in the reliability analysis section * Purpose: to calculate the ICC for reliability studies with multiple quantitative variables * Use-case: in sports science is typically asses the reliability of instruments that report quantitative variables, nowadays we need to compute this calculation using other statistical packages like SPSS.
reli
intercalas correlation coefficient enhancement add the icc calculation in the reliability analysis section purpose to calculate the icc for reliability studies with multiple quantitative variables use case in sports science is typically asses the reliability of instruments that report quantitative variables nowadays we need to compute this calculation using other statistical packages like spss
1
1,581
17,263,685,794
IssuesEvent
2021-07-22 11:07:24
ppy/osu
https://api.github.com/repos/ppy/osu
closed
Game crashes when entering playlists lounge the second time on dev server
type:reliability
**Describe the bug:** 1. Enter the playlists lounge 2. Quit to main menu 3. Enter the playlists lounge again 4. Game crashes when loading the lounge screen for the second time. https://user-images.githubusercontent.com/25472513/126591824-e10a3958-39b9-4c8e-8c5e-3c8dc71fe648.mp4 The dev account I'm using is completely new (just created today, haven't played any maps on it). **osu!lazer version:** master (edce3e0efeaeff1d8479374e834280b2b8bc4c2f) **Logs:** [performance.log](https://github.com/ppy/osu/files/6860086/performance.log) [performance-audio.log](https://github.com/ppy/osu/files/6860087/performance-audio.log) [performance-draw.log](https://github.com/ppy/osu/files/6860088/performance-draw.log) [performance-input.log](https://github.com/ppy/osu/files/6860080/performance-input.log) [performance-update.log](https://github.com/ppy/osu/files/6860081/performance-update.log) [runtime.log](https://github.com/ppy/osu/files/6860082/runtime.log) [updater.log](https://github.com/ppy/osu/files/6860083/updater.log) [database.log](https://github.com/ppy/osu/files/6860084/database.log) [network.log](https://github.com/ppy/osu/files/6860085/network.log)
True
Game crashes when entering playlists lounge the second time on dev server - **Describe the bug:** 1. Enter the playlists lounge 2. Quit to main menu 3. Enter the playlists lounge again 4. Game crashes when loading the lounge screen for the second time. https://user-images.githubusercontent.com/25472513/126591824-e10a3958-39b9-4c8e-8c5e-3c8dc71fe648.mp4 The dev account I'm using is completely new (just created today, haven't played any maps on it). **osu!lazer version:** master (edce3e0efeaeff1d8479374e834280b2b8bc4c2f) **Logs:** [performance.log](https://github.com/ppy/osu/files/6860086/performance.log) [performance-audio.log](https://github.com/ppy/osu/files/6860087/performance-audio.log) [performance-draw.log](https://github.com/ppy/osu/files/6860088/performance-draw.log) [performance-input.log](https://github.com/ppy/osu/files/6860080/performance-input.log) [performance-update.log](https://github.com/ppy/osu/files/6860081/performance-update.log) [runtime.log](https://github.com/ppy/osu/files/6860082/runtime.log) [updater.log](https://github.com/ppy/osu/files/6860083/updater.log) [database.log](https://github.com/ppy/osu/files/6860084/database.log) [network.log](https://github.com/ppy/osu/files/6860085/network.log)
reli
game crashes when entering playlists lounge the second time on dev server describe the bug enter the playlists lounge quit to main menu enter the playlists lounge again game crashes when loading the lounge screen for the second time the dev account i m using is completely new just created today haven t played any maps on it osu lazer version master logs
1
930
11,707,171,881
IssuesEvent
2020-03-08 04:03:17
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Dependency on System.Runtime.InteropServices.RuntimeInformation causes runtime error for apps targeting net451 on *nix systems (i.e. running on Mono)
area-System.Runtime enhancement tenet-reliability untriaged
## Scenario 1) Running a `net451` app depending on `System.Runtime.InteropServices.RuntimeInformation` on OSX/Ubuntu will produce the following exception at runtime: ``` dotnet -v run -f net451 1 ↵ Telemetry is: Enabled Project RuntimeServiceOnMono (.NETFramework,Version=v4.5.1) was previously compiled. Skipping compilation. Running /usr/local/bin/mono --debug /Users/jtluo/Documents/workspace/juntaoluo/tp/RuntimeServiceOnMono/bin/Debug/net451/osx.10.11-x64/RuntimeServiceOnMono.exe Process ID: 29130 Unhandled Exception: System.DllNotFoundException: System.Native at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate () at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0 at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0 at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.DllNotFoundException: System.Native at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate () at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0 at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0 at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0 ``` ## Example See repro at https://github.com/JunTaoLuo/RuntimeServiceOnMono. To run the sample, run `dotnet restore` and `dotnet run -f net451` I understand Mono is not a scenario that's actively being developed for but we should not be causing a exception like this especially since we intend to use the `InteropServices` APIs in often used ASP.NET packages like Logging. Currently this means that most of our samples and apps cannot run on Mono (full CLR on *nix). ## Environment `dotnet --version`: ``` Microsoft .NET Core Shared Framework Host Version : 1.0.1-rc3-004312-00 Build : 1db6c07638a70a621b312e78d4dc9fb31a530f2f ``` `mono -V`: ``` Mono JIT compiler version 4.0.5 ((detached/1d8d582 Tue Oct 20 15:15:33 EDT 2015) Copyright (C) 2002-2014 Novell, Inc, Xamarin Inc and Contributors. www.mono-project.com TLS: normal SIGSEGV: altstack Notification: kqueue Architecture: x86 Disabled: none Misc: softdebug LLVM: yes(3.6.0svn-mono-(detached/a173357) GC: sgen ``` cc @BrennanConroy @muratg @Eilon
True
Dependency on System.Runtime.InteropServices.RuntimeInformation causes runtime error for apps targeting net451 on *nix systems (i.e. running on Mono) - ## Scenario 1) Running a `net451` app depending on `System.Runtime.InteropServices.RuntimeInformation` on OSX/Ubuntu will produce the following exception at runtime: ``` dotnet -v run -f net451 1 ↵ Telemetry is: Enabled Project RuntimeServiceOnMono (.NETFramework,Version=v4.5.1) was previously compiled. Skipping compilation. Running /usr/local/bin/mono --debug /Users/jtluo/Documents/workspace/juntaoluo/tp/RuntimeServiceOnMono/bin/Debug/net451/osx.10.11-x64/RuntimeServiceOnMono.exe Process ID: 29130 Unhandled Exception: System.DllNotFoundException: System.Native at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate () at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0 at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0 at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.DllNotFoundException: System.Native at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate () at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0 at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0 at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0 ``` ## Example See repro at https://github.com/JunTaoLuo/RuntimeServiceOnMono. To run the sample, run `dotnet restore` and `dotnet run -f net451` I understand Mono is not a scenario that's actively being developed for but we should not be causing a exception like this especially since we intend to use the `InteropServices` APIs in often used ASP.NET packages like Logging. Currently this means that most of our samples and apps cannot run on Mono (full CLR on *nix). ## Environment `dotnet --version`: ``` Microsoft .NET Core Shared Framework Host Version : 1.0.1-rc3-004312-00 Build : 1db6c07638a70a621b312e78d4dc9fb31a530f2f ``` `mono -V`: ``` Mono JIT compiler version 4.0.5 ((detached/1d8d582 Tue Oct 20 15:15:33 EDT 2015) Copyright (C) 2002-2014 Novell, Inc, Xamarin Inc and Contributors. www.mono-project.com TLS: normal SIGSEGV: altstack Notification: kqueue Architecture: x86 Disabled: none Misc: softdebug LLVM: yes(3.6.0svn-mono-(detached/a173357) GC: sgen ``` cc @BrennanConroy @muratg @Eilon
reli
dependency on system runtime interopservices runtimeinformation causes runtime error for apps targeting on nix systems i e running on mono scenario running a app depending on system runtime interopservices runtimeinformation on osx ubuntu will produce the following exception at runtime dotnet v run f ↵ telemetry is enabled project runtimeserviceonmono netframework version was previously compiled skipping compilation running usr local bin mono debug users jtluo documents workspace juntaoluo tp runtimeserviceonmono bin debug osx runtimeserviceonmono exe process id unhandled exception system dllnotfoundexception system native at wrapper managed to native interop sys getunixnameprivate at interop sys getunixname in at system runtime interopservices runtimeinformation isosplatform osplatform osplatform in at consoleapplication program main system string args in fatal unhandled exception system dllnotfoundexception system native at wrapper managed to native interop sys getunixnameprivate at interop sys getunixname in at system runtime interopservices runtimeinformation isosplatform osplatform osplatform in at consoleapplication program main system string args in example see repro at to run the sample run dotnet restore and dotnet run f i understand mono is not a scenario that s actively being developed for but we should not be causing a exception like this especially since we intend to use the interopservices apis in often used asp net packages like logging currently this means that most of our samples and apps cannot run on mono full clr on nix environment dotnet version microsoft net core shared framework host version build mono v mono jit compiler version detached tue oct edt copyright c novell inc xamarin inc and contributors tls normal sigsegv altstack notification kqueue architecture disabled none misc softdebug llvm yes mono detached gc sgen cc brennanconroy muratg eilon
1
450
7,612,728,123
IssuesEvent
2018-05-01 18:35:57
m3db/m3db
https://api.github.com/repos/m3db/m3db
opened
Monitor file sizes
C: Toolchain P: Medium T: Instrumentation T: Reliability
Add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard/index level. i.e. ``` /data/namespace/shard <-- data fileset files /snapshot/namespace/shard <-- snapshot fileset files ```
True
Monitor file sizes - Add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard/index level. i.e. ``` /data/namespace/shard <-- data fileset files /snapshot/namespace/shard <-- snapshot fileset files ```
reli
monitor file sizes add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard index level i e data namespace shard data fileset files snapshot namespace shard snapshot fileset files
1
271,689
8,488,531,242
IssuesEvent
2018-10-26 16:56:07
FIDUCEO/FCDR_HIRS
https://api.github.com/repos/FIDUCEO/FCDR_HIRS
closed
Odd values for channel 4 and 12 on metopa
Priority: High
![image](https://user-images.githubusercontent.com/500246/45166381-c6c60880-b1ee-11e8-9b1c-6fec62d01351.png) particularly clear in time series ![image](https://user-images.githubusercontent.com/500246/45168826-24a91f00-b1f4-11e8-973b-51da94e557d6.png) should investigate! Compare also Shi and Bates (2011).
1.0
Odd values for channel 4 and 12 on metopa - ![image](https://user-images.githubusercontent.com/500246/45166381-c6c60880-b1ee-11e8-9b1c-6fec62d01351.png) particularly clear in time series ![image](https://user-images.githubusercontent.com/500246/45168826-24a91f00-b1f4-11e8-973b-51da94e557d6.png) should investigate! Compare also Shi and Bates (2011).
non_reli
odd values for channel and on metopa particularly clear in time series should investigate compare also shi and bates
0
35,914
2,793,820,595
IssuesEvent
2015-05-11 13:37:31
elecoest/allevents-3-2
https://api.github.com/repos/elecoest/allevents-3-2
closed
FrontEnd - Amélioration - Rajout page affichage des lieux et événements géolocalisés
auto-migrated Priority-Medium Type-Enhancement
``` Dans AE on dispose d'une liste de lieux qu'il serait intéressant de pouvoir visualiser en front sous forme d'une carte. En effet, les lieux sont un peu ce qui différencie AE de certaines autres extensions de gestion d'agendas qui n'en possèdent pas. Pouvoir les mettre en valeur (et les utiliser) en front me semble couler de source. Il existe bien des composants spécialisés dans l'affichage sous forme de carte de lieux ou de contacts mais si on voulait les utiliser avec AE, il faudrait alors re-saisir les lieux (latitude, longitude, titres ...) dans ces composants ce qui me semble un peu dommage. Dans AE la page d'affichage de carte pourrait avoir plusieurs modes d'affichages : - Carte des Lieux - Carte des événements (par lieu et par date). Avec filtres sur agendas, catégories etc... en fonction du choix et des possibilités de filtrage des éléments choisis. Sur la carte à la place d'une icône par défaut, les vignettes pourraient être utilisées. Au survol de chaque éléments, une fenêtre de type Pop-up pourrait afficher le détail de l'élément survolé. Si une adresse mail est disponible, un clic sur le titre de l'élément (dans la fenêtre pop-up) permettrait d'écrire au responsable du lieux ou de l'événement (choix de permettre ou non cette fonction dans les paramètres de la page). Un exemple ci-dessous. ``` Original issue reported on code.google.com by `jjacquesh` on 13 Dec 2014 at 11:20 Attachments: * [20141213_cartographie_des_lieux.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-302/comment-0/20141213_cartographie_des_lieux.jpg)
1.0
FrontEnd - Amélioration - Rajout page affichage des lieux et événements géolocalisés - ``` Dans AE on dispose d'une liste de lieux qu'il serait intéressant de pouvoir visualiser en front sous forme d'une carte. En effet, les lieux sont un peu ce qui différencie AE de certaines autres extensions de gestion d'agendas qui n'en possèdent pas. Pouvoir les mettre en valeur (et les utiliser) en front me semble couler de source. Il existe bien des composants spécialisés dans l'affichage sous forme de carte de lieux ou de contacts mais si on voulait les utiliser avec AE, il faudrait alors re-saisir les lieux (latitude, longitude, titres ...) dans ces composants ce qui me semble un peu dommage. Dans AE la page d'affichage de carte pourrait avoir plusieurs modes d'affichages : - Carte des Lieux - Carte des événements (par lieu et par date). Avec filtres sur agendas, catégories etc... en fonction du choix et des possibilités de filtrage des éléments choisis. Sur la carte à la place d'une icône par défaut, les vignettes pourraient être utilisées. Au survol de chaque éléments, une fenêtre de type Pop-up pourrait afficher le détail de l'élément survolé. Si une adresse mail est disponible, un clic sur le titre de l'élément (dans la fenêtre pop-up) permettrait d'écrire au responsable du lieux ou de l'événement (choix de permettre ou non cette fonction dans les paramètres de la page). Un exemple ci-dessous. ``` Original issue reported on code.google.com by `jjacquesh` on 13 Dec 2014 at 11:20 Attachments: * [20141213_cartographie_des_lieux.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-302/comment-0/20141213_cartographie_des_lieux.jpg)
non_reli
frontend amélioration rajout page affichage des lieux et événements géolocalisés dans ae on dispose d une liste de lieux qu il serait intéressant de pouvoir visualiser en front sous forme d une carte en effet les lieux sont un peu ce qui différencie ae de certaines autres extensions de gestion d agendas qui n en possèdent pas pouvoir les mettre en valeur et les utiliser en front me semble couler de source il existe bien des composants spécialisés dans l affichage sous forme de carte de lieux ou de contacts mais si on voulait les utiliser avec ae il faudrait alors re saisir les lieux latitude longitude titres dans ces composants ce qui me semble un peu dommage dans ae la page d affichage de carte pourrait avoir plusieurs modes d affichages carte des lieux carte des événements par lieu et par date avec filtres sur agendas catégories etc en fonction du choix et des possibilités de filtrage des éléments choisis sur la carte à la place d une icône par défaut les vignettes pourraient être utilisées au survol de chaque éléments une fenêtre de type pop up pourrait afficher le détail de l élément survolé si une adresse mail est disponible un clic sur le titre de l élément dans la fenêtre pop up permettrait d écrire au responsable du lieux ou de l événement choix de permettre ou non cette fonction dans les paramètres de la page un exemple ci dessous original issue reported on code google com by jjacquesh on dec at attachments
0
2,489
25,770,265,519
IssuesEvent
2022-12-09 07:15:21
juliuspf/jasp-issues
https://api.github.com/repos/juliuspf/jasp-issues
closed
oh new please
Module: jaspReliability Feature Request
### Description _No response_ ### Purpose _No response_ ### Use-case _No response_ ### Is your feature request related to a problem? _No response_ ### Is your feature request related to a JASP module? Reliability ### Describe the solution you would like _No response_ ### Describe alternatives that you have considered _No response_ ### Additional context _No response_
True
oh new please - ### Description _No response_ ### Purpose _No response_ ### Use-case _No response_ ### Is your feature request related to a problem? _No response_ ### Is your feature request related to a JASP module? Reliability ### Describe the solution you would like _No response_ ### Describe alternatives that you have considered _No response_ ### Additional context _No response_
reli
oh new please description no response purpose no response use case no response is your feature request related to a problem no response is your feature request related to a jasp module reliability describe the solution you would like no response describe alternatives that you have considered no response additional context no response
1
1,682
18,518,547,317
IssuesEvent
2021-10-20 12:55:00
hashicorp/consul
https://api.github.com/repos/hashicorp/consul
closed
docs: Note the Grafana dashboard
type/docs good first issue theme/telemetry theme/reliability Hacktoberfest
Since last year we've had an "official" Grafana dashboard which lives in the source tree and we [publish publicly](https://grafana.com/grafana/dashboards/13396). But many users including internal teams aren't aware it exists. We should update the Telemetry page to note it's existence!
True
docs: Note the Grafana dashboard - Since last year we've had an "official" Grafana dashboard which lives in the source tree and we [publish publicly](https://grafana.com/grafana/dashboards/13396). But many users including internal teams aren't aware it exists. We should update the Telemetry page to note it's existence!
reli
docs note the grafana dashboard since last year we ve had an official grafana dashboard which lives in the source tree and we but many users including internal teams aren t aware it exists we should update the telemetry page to note it s existence
1
1,532
16,759,295,826
IssuesEvent
2021-06-13 13:01:35
skytable/skytable
https://api.github.com/repos/skytable/skytable
closed
Server: Running two skyd in same directory can lead to loss of data
A-independent C-bug C-reliability C-storage D-server P-high S-waiting-on-author
**Description of bug** Currently it is possible to start two `skyd` processes from the same directory. This will lead to data being overwritten by the other process. **Steps to reproduce** Steps to reproduce the behavior: For example: 1. Run `skyd` 2. Run `skyd` again 3. Store some data on both servers 3. When both `skyd`are closed you will end up with data for only the last server closed **Expected behavior** Stop any subsequent skyd process from starting up if there is already one running in that directory. **Meta** - **Release tag:** None - **Branch:** - **Commit ID:** - **Operating system:**
True
Server: Running two skyd in same directory can lead to loss of data - **Description of bug** Currently it is possible to start two `skyd` processes from the same directory. This will lead to data being overwritten by the other process. **Steps to reproduce** Steps to reproduce the behavior: For example: 1. Run `skyd` 2. Run `skyd` again 3. Store some data on both servers 3. When both `skyd`are closed you will end up with data for only the last server closed **Expected behavior** Stop any subsequent skyd process from starting up if there is already one running in that directory. **Meta** - **Release tag:** None - **Branch:** - **Commit ID:** - **Operating system:**
reli
server running two skyd in same directory can lead to loss of data description of bug currently it is possible to start two skyd processes from the same directory this will lead to data being overwritten by the other process steps to reproduce steps to reproduce the behavior for example run skyd run skyd again store some data on both servers when both skyd are closed you will end up with data for only the last server closed expected behavior stop any subsequent skyd process from starting up if there is already one running in that directory meta release tag none branch commit id operating system
1
1,799
19,915,685,483
IssuesEvent
2022-01-25 22:19:05
Azure/azure-sdk-for-java
https://api.github.com/repos/Azure/azure-sdk-for-java
closed
DownloadToFile causes an error - reactor.core.Exceptions$OverflowException & java.util.concurrent.TimeoutException
question Storage Client customer-reported pillar-reliability
2021-10-21 16:34:04,381 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:04,383 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:181) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:04,410 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:20,566 ERROR [Thread-55] (AzureDownloadToLocalFileCopy:62) - Azure copyFile failed : java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) reactor.core.Exceptions$ReactiveException: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) at reactor.core.Exceptions.propagate(Exceptions.java:392) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:97) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?] at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?] at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?] at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?] at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?] at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?] Suppressed: java.lang.Exception: #block terminated with an error at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?] at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?] at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?] at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?] at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?] at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?] Caused by: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.handleTimeout(FluxTimeout.java:294) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.doTimeout(FluxTimeout.java:279) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutTimeoutSubscriber.onNext(FluxTimeout.java:418) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.MonoDelay$MonoDelayRunnable.propagateDelay(MonoDelay.java:270) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.MonoDelay$MonoDelayRunnable.run(MonoDelay.java:285) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28) ~[reactor-core-3.4.6.jar!/:3.4.6] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?] at java.lang.Thread.run(Thread.java:834) ~[?:?]
True
DownloadToFile causes an error - reactor.core.Exceptions$OverflowException & java.util.concurrent.TimeoutException - 2021-10-21 16:34:04,381 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:04,383 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:181) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:04,410 DEBUG [Thread-64] (Operators:254) - More data produced than requested reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...) at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?] at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?] at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?] at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?] at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] 2021-10-21 16:34:20,566 ERROR [Thread-55] (AzureDownloadToLocalFileCopy:62) - Azure copyFile failed : java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) reactor.core.Exceptions$ReactiveException: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) at reactor.core.Exceptions.propagate(Exceptions.java:392) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:97) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?] at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?] at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?] at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?] at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?] at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?] Suppressed: java.lang.Exception: #block terminated with an error at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6] at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?] at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?] at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?] at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?] at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?] at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?] at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?] at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?] at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?] Caused by: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured) at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.handleTimeout(FluxTimeout.java:294) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.doTimeout(FluxTimeout.java:279) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxTimeout$TimeoutTimeoutSubscriber.onNext(FluxTimeout.java:418) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.MonoDelay$MonoDelayRunnable.propagateDelay(MonoDelay.java:270) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.publisher.MonoDelay$MonoDelayRunnable.run(MonoDelay.java:285) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68) ~[reactor-core-3.4.6.jar!/:3.4.6] at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28) ~[reactor-core-3.4.6.jar!/:3.4.6] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?] at java.lang.Thread.run(Thread.java:834) ~[?:?]
reli
downloadtofile causes an error reactor core exceptions overflowexception java util concurrent timeoutexception debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmap mapsubscriber onnext fluxmap java at reactor core publisher serializedsubscriber onnext serializedsubscriber java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmap mapsubscriber onnext fluxmap java at reactor core publisher serializedsubscriber onnext serializedsubscriber java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java error azuredownloadtolocalfilecopy azure copyfile failed java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured reactor core exceptions reactiveexception java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured at reactor core exceptions propagate exceptions java at reactor core publisher blockingsinglesubscriber blockingget blockingsinglesubscriber java at reactor core publisher mono block mono java at com azure storage common implementation storageimplutils blockwithoptionaltimeout storageimplutils java at com azure storage blob specialized blobclientbase downloadtofilewithresponse blobclientbase java at com bmc cm aft client wrapper azurewrapper downloadblob azurewrapper java at com bmc cm aft client azuredownloadtolocalfilecopy copyfile azuredownloadtolocalfilecopy java at com bmc cm aft client filecopy transfer filecopy java at com bmc cm aft client filecopy run filecopy java at com bmc cm aft client ftctransfer copysinglefile ftctransfer java at com bmc cm aft client ftctransfer copy ftctransfer java at com bmc cm aft client ftctransfer runcopy ftctransfer java at com bmc cm aft client ftctransfer run ftctransfer java at com bmc cm aft client subtaskrunner run subtaskrunner java at com bmc cm aft client ftchandler runsubtasks ftchandler java at com bmc cm aft client ftchandler runjob ftchandler java at com bmc cm aft jobmanagement jobobjectimpl jobrunnerthread run jobobjectimpl java suppressed java lang exception block terminated with an error at reactor core publisher blockingsinglesubscriber blockingget blockingsinglesubscriber java at reactor core publisher mono block mono java at com azure storage common implementation storageimplutils blockwithoptionaltimeout storageimplutils java at com azure storage blob specialized blobclientbase downloadtofilewithresponse blobclientbase java at com bmc cm aft client wrapper azurewrapper downloadblob azurewrapper java at com bmc cm aft client azuredownloadtolocalfilecopy copyfile azuredownloadtolocalfilecopy java at com bmc cm aft client filecopy transfer filecopy java at com bmc cm aft client filecopy run filecopy java at com bmc cm aft client ftctransfer copysinglefile ftctransfer java at com bmc cm aft client ftctransfer copy ftctransfer java at com bmc cm aft client ftctransfer runcopy ftctransfer java at com bmc cm aft client ftctransfer run ftctransfer java at com bmc cm aft client subtaskrunner run subtaskrunner java at com bmc cm aft client ftchandler runsubtasks ftchandler java at com bmc cm aft client ftchandler runjob ftchandler java at com bmc cm aft jobmanagement jobobjectimpl jobrunnerthread run jobobjectimpl java caused by java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured at reactor core publisher fluxtimeout timeoutmainsubscriber handletimeout fluxtimeout java at reactor core publisher fluxtimeout timeoutmainsubscriber dotimeout fluxtimeout java at reactor core publisher fluxtimeout timeouttimeoutsubscriber onnext fluxtimeout java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher monodelay monodelayrunnable propagatedelay monodelay java at reactor core publisher monodelay monodelayrunnable run monodelay java at reactor core scheduler schedulertask call schedulertask java at reactor core scheduler schedulertask call schedulertask java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
1
1,970
22,293,542,308
IssuesEvent
2022-06-12 18:13:53
jina-ai/jina
https://api.github.com/repos/jina-ai/jina
closed
Server side throttling
epic/reliability
At server side throttling so that Gateway/Head/Worker can reject new requests if there are too many outstanding requests already queued up. This needs to configurable
True
Server side throttling - At server side throttling so that Gateway/Head/Worker can reject new requests if there are too many outstanding requests already queued up. This needs to configurable
reli
server side throttling at server side throttling so that gateway head worker can reject new requests if there are too many outstanding requests already queued up this needs to configurable
1
2,390
25,100,822,486
IssuesEvent
2022-11-08 13:31:00
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
opened
Disable the message TTL by default
kind/toil scope/clients-java scope/clients-go severity/mid area/reliability
**Description** I can publish a message using Zeebe's Java or Go client. The message can have a time-to-live (TTL). It can be correlated within the given TTL. After the TTL is reached, the message expires and is removed. Currently, the Java and the Go client use a different default value for the message TTL. * in the Java client: the default TTL is [1 hour](https://github.com/camunda/zeebe/blob/main/clients/java/src/main/java/io/camunda/zeebe/client/impl/ZeebeClientBuilderImpl.java#L57) * in the Go client: the default TTL is [5 seconds](https://github.com/camunda/zeebe/blob/main/clients/go/cmd/zbctl/internal/commands/publishMessage.go#L74) We should align the default value in all (official) clients to align the general behavior. I recommend setting the default message TTL to `zero` (i.e. disable the message buffering). If the TTL is set to a high value then it can impact the performance of the broker (see https://github.com/camunda/zeebe/issues/10643#issuecomment-1303710858). Instead of setting an arbitrary value for the TTL, let the user explicitly choose to use the buffering and think about a reasonable value.
True
Disable the message TTL by default - **Description** I can publish a message using Zeebe's Java or Go client. The message can have a time-to-live (TTL). It can be correlated within the given TTL. After the TTL is reached, the message expires and is removed. Currently, the Java and the Go client use a different default value for the message TTL. * in the Java client: the default TTL is [1 hour](https://github.com/camunda/zeebe/blob/main/clients/java/src/main/java/io/camunda/zeebe/client/impl/ZeebeClientBuilderImpl.java#L57) * in the Go client: the default TTL is [5 seconds](https://github.com/camunda/zeebe/blob/main/clients/go/cmd/zbctl/internal/commands/publishMessage.go#L74) We should align the default value in all (official) clients to align the general behavior. I recommend setting the default message TTL to `zero` (i.e. disable the message buffering). If the TTL is set to a high value then it can impact the performance of the broker (see https://github.com/camunda/zeebe/issues/10643#issuecomment-1303710858). Instead of setting an arbitrary value for the TTL, let the user explicitly choose to use the buffering and think about a reasonable value.
reli
disable the message ttl by default description i can publish a message using zeebe s java or go client the message can have a time to live ttl it can be correlated within the given ttl after the ttl is reached the message expires and is removed currently the java and the go client use a different default value for the message ttl in the java client the default ttl is in the go client the default ttl is we should align the default value in all official clients to align the general behavior i recommend setting the default message ttl to zero i e disable the message buffering if the ttl is set to a high value then it can impact the performance of the broker see instead of setting an arbitrary value for the ttl let the user explicitly choose to use the buffering and think about a reasonable value
1
306
6,396,443,449
IssuesEvent
2017-08-04 15:33:40
rook/rook
https://api.github.com/repos/rook/rook
closed
Cluster fails to come up after a reboot
rados reliability
The mons are dependent on a constant network identity. If that identity changes, nobody will be able to connect to the mon. The mon will be rejected. Ceph enforces that the mon endpoint is what the mon binds to. Putting a k8s service in front of a mon to have a consistent endpoint does not work. Currently the mons use the PodIP, which will change at every restart. We must either find a way to soften the ceph endpoint requirements to allow the service translation, or find a way to get a consistent endpoint for everyone to talk to the mons, even after reboot.
True
Cluster fails to come up after a reboot - The mons are dependent on a constant network identity. If that identity changes, nobody will be able to connect to the mon. The mon will be rejected. Ceph enforces that the mon endpoint is what the mon binds to. Putting a k8s service in front of a mon to have a consistent endpoint does not work. Currently the mons use the PodIP, which will change at every restart. We must either find a way to soften the ceph endpoint requirements to allow the service translation, or find a way to get a consistent endpoint for everyone to talk to the mons, even after reboot.
reli
cluster fails to come up after a reboot the mons are dependent on a constant network identity if that identity changes nobody will be able to connect to the mon the mon will be rejected ceph enforces that the mon endpoint is what the mon binds to putting a service in front of a mon to have a consistent endpoint does not work currently the mons use the podip which will change at every restart we must either find a way to soften the ceph endpoint requirements to allow the service translation or find a way to get a consistent endpoint for everyone to talk to the mons even after reboot
1
156,814
5,989,608,736
IssuesEvent
2017-06-02 09:04:52
universAAL/ui
https://api.github.com/repos/universAAL/ui
closed
UI frozen
bug imported priority 3
_Originally Opened: @dimokas (2013-03-22 16:02:30_) _Originally Closed: 2013-04-22 10:56:24_ In some cases, when the service call takes some time (few seconds or less) then the UI is not repsonding. The user cannot press any button since there is no interaction. It is like a frozen window. -- From: _this issue has been automatically imported from our old issue tracker_
1.0
UI frozen - _Originally Opened: @dimokas (2013-03-22 16:02:30_) _Originally Closed: 2013-04-22 10:56:24_ In some cases, when the service call takes some time (few seconds or less) then the UI is not repsonding. The user cannot press any button since there is no interaction. It is like a frozen window. -- From: _this issue has been automatically imported from our old issue tracker_
non_reli
ui frozen originally opened dimokas originally closed in some cases when the service call takes some time few seconds or less then the ui is not repsonding the user cannot press any button since there is no interaction it is like a frozen window from this issue has been automatically imported from our old issue tracker
0
120,573
4,791,766,652
IssuesEvent
2016-10-31 13:44:07
vidalborromeo/eecshelp-beta
https://api.github.com/repos/vidalborromeo/eecshelp-beta
opened
For Moderator Signed-In On Multiple Browsers, First Sign-Off Makes Them Appear "Offline"
Component: Backend Priority: Low Type: Bug
If a moderator is signed in via multiple browsers, the first Unsubscribe event will make it appear that they are no longer online. Solution: During the backend query in Unsubscribe, check if there are other Observers associated with that User entity.
1.0
For Moderator Signed-In On Multiple Browsers, First Sign-Off Makes Them Appear "Offline" - If a moderator is signed in via multiple browsers, the first Unsubscribe event will make it appear that they are no longer online. Solution: During the backend query in Unsubscribe, check if there are other Observers associated with that User entity.
non_reli
for moderator signed in on multiple browsers first sign off makes them appear offline if a moderator is signed in via multiple browsers the first unsubscribe event will make it appear that they are no longer online solution during the backend query in unsubscribe check if there are other observers associated with that user entity
0
352,589
32,078,556,914
IssuesEvent
2023-09-25 12:37:39
durof/status
https://api.github.com/repos/durof/status
closed
🛑 Mention Test Bot is down
status mention-test-bot
In [`00c5ad2`](https://github.com/durof/status/commit/00c5ad28047456252721048b7f545d3ffa7c9799 ), Mention Test Bot (https://backend.isbotdown.com/bots/mentiontestbot) was **down**: - HTTP code: 200 - Response time: 88 ms
1.0
🛑 Mention Test Bot is down - In [`00c5ad2`](https://github.com/durof/status/commit/00c5ad28047456252721048b7f545d3ffa7c9799 ), Mention Test Bot (https://backend.isbotdown.com/bots/mentiontestbot) was **down**: - HTTP code: 200 - Response time: 88 ms
non_reli
🛑 mention test bot is down in mention test bot was down http code response time ms
0
1,077
12,885,581,771
IssuesEvent
2020-07-13 08:00:25
Azure/azure-sdk-for-java
https://api.github.com/repos/Azure/azure-sdk-for-java
closed
[BUG] Interminent and recurring JsonParseException
Compute - VM Mgmt customer-reported needs-author-feedback no-recent-activity tenet-reliability
**Describe the bug** Often when using the sdk or even plain http requests to fetch information from azure, an unexpected response is given. The response looks like html code, rather then JSON. ***Exception or Stack Trace*** `java.lang.RuntimeException: com.fasterxml.jackson.core.JsonParseException: Unexpected character ('<' (code 60)): expected a valid value (number, String, array, object, 'true', 'false' or 'null') at [Source: (String)"<!doctype html> <html> <head> <meta charset="utf-8"> <meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1"> <title>Symantec CloudSOC™</title> <meta name="robots" content="noindex, nofollow" /> <meta name="description" content=""> <meta http-equiv="Content-Security-Policy" content="default-src * data: blob:;script-src *.elastica.net *.elastica-inc.com *.nr-data.net *.newrelic.com 'unsafe-inline' 'unsafe-eval' blob: data: 'self';style-src data: blob: 'unsafe-inline'"[truncated 1533 chars]; line: 1, column: 2] at rx.exceptions.Exceptions.propagate(Exceptions.java:57) at rx.observables.BlockingObservable.forEach(BlockingObservable.java:139) ` **To Reproduce** Steps to reproduce the behavior: I am not sure but to me, I just need to execute any sdk list operation. ***Code Snippet*** azure.virtualMachines().listByResouceGroup And many others. **Expected behavior** Every response should be a valid json. **Setup (please complete the following information):** - OS: windows - IDE : intelliJ - Version of the Library used 1.31.1 **Information Checklist** Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report - [ x] Bug Description Added - [ x] Repro Steps Added - [x ] Setup information Added
True
[BUG] Interminent and recurring JsonParseException - **Describe the bug** Often when using the sdk or even plain http requests to fetch information from azure, an unexpected response is given. The response looks like html code, rather then JSON. ***Exception or Stack Trace*** `java.lang.RuntimeException: com.fasterxml.jackson.core.JsonParseException: Unexpected character ('<' (code 60)): expected a valid value (number, String, array, object, 'true', 'false' or 'null') at [Source: (String)"<!doctype html> <html> <head> <meta charset="utf-8"> <meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1"> <title>Symantec CloudSOC™</title> <meta name="robots" content="noindex, nofollow" /> <meta name="description" content=""> <meta http-equiv="Content-Security-Policy" content="default-src * data: blob:;script-src *.elastica.net *.elastica-inc.com *.nr-data.net *.newrelic.com 'unsafe-inline' 'unsafe-eval' blob: data: 'self';style-src data: blob: 'unsafe-inline'"[truncated 1533 chars]; line: 1, column: 2] at rx.exceptions.Exceptions.propagate(Exceptions.java:57) at rx.observables.BlockingObservable.forEach(BlockingObservable.java:139) ` **To Reproduce** Steps to reproduce the behavior: I am not sure but to me, I just need to execute any sdk list operation. ***Code Snippet*** azure.virtualMachines().listByResouceGroup And many others. **Expected behavior** Every response should be a valid json. **Setup (please complete the following information):** - OS: windows - IDE : intelliJ - Version of the Library used 1.31.1 **Information Checklist** Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report - [ x] Bug Description Added - [ x] Repro Steps Added - [x ] Setup information Added
reli
interminent and recurring jsonparseexception describe the bug often when using the sdk or even plain http requests to fetch information from azure an unexpected response is given the response looks like html code rather then json exception or stack trace java lang runtimeexception com fasterxml jackson core jsonparseexception unexpected character code expected a valid value number string array object true false or null at source string symantec cloudsoc™ meta http equiv content security policy content default src data blob script src elastica net elastica inc com nr data net newrelic com unsafe inline unsafe eval blob data self style src data blob unsafe inline line column at rx exceptions exceptions propagate exceptions java at rx observables blockingobservable foreach blockingobservable java to reproduce steps to reproduce the behavior i am not sure but to me i just need to execute any sdk list operation code snippet azure virtualmachines listbyresoucegroup and many others expected behavior every response should be a valid json setup please complete the following information os windows ide intellij version of the library used information checklist kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report bug description added repro steps added setup information added
1
389,614
26,825,331,072
IssuesEvent
2023-02-02 12:28:20
nextauthjs/next-auth
https://api.github.com/repos/nextauthjs/next-auth
closed
Fix prisma doc schema.prisma userId warning for mongodb
documentation triage
### What is the improvement or update you wish to see? Don't show the warning message. **"Warning validating field user in model Account: Field Account.userId and User.id must have the same native type for MongoDB to join those collections correctly. Consider either removing User.id's native type attribute or adding '@db.ObjectId' to Account.userId. Beware that this will become an error in the future."** ### Is there any context that might help us understand? Prisma member recommended add `@db.ObjectId` for `userId` field to fix this problem. [link](https://github.com/prisma/prisma/discussions/17624#discussioncomment-4820223) ### Does the docs page already exist? Please link to it. https://next-auth.js.org/adapters/prisma
1.0
Fix prisma doc schema.prisma userId warning for mongodb - ### What is the improvement or update you wish to see? Don't show the warning message. **"Warning validating field user in model Account: Field Account.userId and User.id must have the same native type for MongoDB to join those collections correctly. Consider either removing User.id's native type attribute or adding '@db.ObjectId' to Account.userId. Beware that this will become an error in the future."** ### Is there any context that might help us understand? Prisma member recommended add `@db.ObjectId` for `userId` field to fix this problem. [link](https://github.com/prisma/prisma/discussions/17624#discussioncomment-4820223) ### Does the docs page already exist? Please link to it. https://next-auth.js.org/adapters/prisma
non_reli
fix prisma doc schema prisma userid warning for mongodb what is the improvement or update you wish to see don t show the warning message warning validating field user in model account field account userid and user id must have the same native type for mongodb to join those collections correctly consider either removing user id s native type attribute or adding db objectid to account userid beware that this will become an error in the future is there any context that might help us understand prisma member recommended add db objectid for userid field to fix this problem does the docs page already exist please link to it
0
303,274
22,961,863,212
IssuesEvent
2022-07-19 16:00:15
Workday/canvas-kit
https://api.github.com/repos/Workday/canvas-kit
closed
Update ActionBar section in V7 upgrade guide
documentation s:1 7.x
## 💡 Suggestion The ActionBar section in the V7 upgrade guide could use a bit more detail on the removal of the `fixed` prop. ## Motivation There have been a few support questions that have come up recently.
1.0
Update ActionBar section in V7 upgrade guide - ## 💡 Suggestion The ActionBar section in the V7 upgrade guide could use a bit more detail on the removal of the `fixed` prop. ## Motivation There have been a few support questions that have come up recently.
non_reli
update actionbar section in upgrade guide 💡 suggestion the actionbar section in the upgrade guide could use a bit more detail on the removal of the fixed prop motivation there have been a few support questions that have come up recently
0
46,631
13,055,950,338
IssuesEvent
2020-07-30 03:12:43
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[docs] Broken on combo trunk (Trac #1613)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/1613 ```json { "status": "closed", "changetime": "2016-04-05T23:31:03", "description": "This is the error I get on combo trunk Ubuntu 14.04 with sphinx 1.4 (recently updated 'cause the napolean extension is now required).\n\nException occurred:\n File \"/usr/local/lib/python2.7/dist-packages/sphinx/domains/cpp.py\", line 4039, in _resolve_xref_inner\n assert parentSymbol # should be there\nAssertionError\n", "reporter": "olivas", "cc": "", "resolution": "worksforme", "_ts": "1459899063339623", "component": "cmake", "summary": "[docs] Broken on combo trunk", "priority": "major", "keywords": "", "time": "2016-03-30T13:43:09", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
[docs] Broken on combo trunk (Trac #1613) - Migrated from https://code.icecube.wisc.edu/ticket/1613 ```json { "status": "closed", "changetime": "2016-04-05T23:31:03", "description": "This is the error I get on combo trunk Ubuntu 14.04 with sphinx 1.4 (recently updated 'cause the napolean extension is now required).\n\nException occurred:\n File \"/usr/local/lib/python2.7/dist-packages/sphinx/domains/cpp.py\", line 4039, in _resolve_xref_inner\n assert parentSymbol # should be there\nAssertionError\n", "reporter": "olivas", "cc": "", "resolution": "worksforme", "_ts": "1459899063339623", "component": "cmake", "summary": "[docs] Broken on combo trunk", "priority": "major", "keywords": "", "time": "2016-03-30T13:43:09", "milestone": "", "owner": "nega", "type": "defect" } ```
non_reli
broken on combo trunk trac migrated from json status closed changetime description this is the error i get on combo trunk ubuntu with sphinx recently updated cause the napolean extension is now required n nexception occurred n file usr local lib dist packages sphinx domains cpp py line in resolve xref inner n assert parentsymbol should be there nassertionerror n reporter olivas cc resolution worksforme ts component cmake summary broken on combo trunk priority major keywords time milestone owner nega type defect
0
697
9,934,215,224
IssuesEvent
2019-07-02 14:02:09
plotly/dash-table
https://api.github.com/repos/plotly/dash-table
closed
add test for semi editable table
Attribute: Reliability
verify setting editable=True at table level and editable=False at column level to disable editing of specified columns works.
True
add test for semi editable table - verify setting editable=True at table level and editable=False at column level to disable editing of specified columns works.
reli
add test for semi editable table verify setting editable true at table level and editable false at column level to disable editing of specified columns works
1
780,308
27,389,105,401
IssuesEvent
2023-02-28 15:11:07
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
GOCA: Vm migration error -501
Status: Pending Type: Bug Category: API Priority: Normal
OpenNebula 6.0.0.2 github.com/OpenNebula/one/src/oca/go/src/goca v0.0.0-20230130124959-3c9bfecbd197 Try if err := n.controller.VM(134).Migrate(1, true, true, 0, 0); err != nil { return err } but have err: Code: -501 String: "Parameter that is supposed to be integer is not"
1.0
GOCA: Vm migration error -501 - OpenNebula 6.0.0.2 github.com/OpenNebula/one/src/oca/go/src/goca v0.0.0-20230130124959-3c9bfecbd197 Try if err := n.controller.VM(134).Migrate(1, true, true, 0, 0); err != nil { return err } but have err: Code: -501 String: "Parameter that is supposed to be integer is not"
non_reli
goca vm migration error opennebula github com opennebula one src oca go src goca try if err n controller vm migrate true true err nil return err but have err code string parameter that is supposed to be integer is not
0
1,185
13,634,351,343
IssuesEvent
2020-09-24 23:35:50
Azure/azure-sdk-for-java
https://api.github.com/repos/Azure/azure-sdk-for-java
opened
Ensure all Key Vault models align correctly with the Swagger specification
Client KeyVault tenet-reliability
Issue #14513 showed that the Key Vault Keys library had issues when deserializing a couple models (fixed by #15375), so it would be a worthwhile effort to further look into the rest of the models in Keys, Certificates and Secrets to ensure everything conforms to the service's Swagger specification.
True
Ensure all Key Vault models align correctly with the Swagger specification - Issue #14513 showed that the Key Vault Keys library had issues when deserializing a couple models (fixed by #15375), so it would be a worthwhile effort to further look into the rest of the models in Keys, Certificates and Secrets to ensure everything conforms to the service's Swagger specification.
reli
ensure all key vault models align correctly with the swagger specification issue showed that the key vault keys library had issues when deserializing a couple models fixed by so it would be a worthwhile effort to further look into the rest of the models in keys certificates and secrets to ensure everything conforms to the service s swagger specification
1
77
3,477,830,527
IssuesEvent
2015-12-28 06:21:47
PopupMaker/Popup-Maker
https://api.github.com/repos/PopupMaker/Popup-Maker
closed
Introduce new Popup class.
Reliability
This will simplify the data model & methods needed for both standard template loading and WP API loading. Methods is_loadable() title() content() theme_id() close_text()
True
Introduce new Popup class. - This will simplify the data model & methods needed for both standard template loading and WP API loading. Methods is_loadable() title() content() theme_id() close_text()
reli
introduce new popup class this will simplify the data model methods needed for both standard template loading and wp api loading methods is loadable title content theme id close text
1
278,999
21,104,597,275
IssuesEvent
2022-04-04 17:27:33
WaffleKone/Blades-Companion
https://api.github.com/repos/WaffleKone/Blades-Companion
opened
Configure boiler plates and linting
documentation
Make it look pretty and good! - [ ] Set up and configure ESLint and Prettier and make them work together - [ ] Set up automatic testing and changes with ESLint when commits are made - [ ] Set up all dependencies in package.json
1.0
Configure boiler plates and linting - Make it look pretty and good! - [ ] Set up and configure ESLint and Prettier and make them work together - [ ] Set up automatic testing and changes with ESLint when commits are made - [ ] Set up all dependencies in package.json
non_reli
configure boiler plates and linting make it look pretty and good set up and configure eslint and prettier and make them work together set up automatic testing and changes with eslint when commits are made set up all dependencies in package json
0
300
6,150,716,165
IssuesEvent
2017-06-27 23:31:29
CCI-MOC/hil
https://api.github.com/repos/CCI-MOC/hil
closed
Using SNMP to control switches to set VLANS
Reliability
We should have a better way to set VLANS programmatically than using pexpect due to the problems discussed in #744. @pjd-nu gave us super overview of how SNMP works and what troubles we may encounter (like it is good for reading stuff from hardware, but writing to hardware can be tedious at times) I, @henn, @naved001, @pjd-nu agree that this can be a good internship project. Explore, understand SNMP and come up with a solution to set VLANS using it.
True
Using SNMP to control switches to set VLANS - We should have a better way to set VLANS programmatically than using pexpect due to the problems discussed in #744. @pjd-nu gave us super overview of how SNMP works and what troubles we may encounter (like it is good for reading stuff from hardware, but writing to hardware can be tedious at times) I, @henn, @naved001, @pjd-nu agree that this can be a good internship project. Explore, understand SNMP and come up with a solution to set VLANS using it.
reli
using snmp to control switches to set vlans we should have a better way to set vlans programmatically than using pexpect due to the problems discussed in pjd nu gave us super overview of how snmp works and what troubles we may encounter like it is good for reading stuff from hardware but writing to hardware can be tedious at times i henn pjd nu agree that this can be a good internship project explore understand snmp and come up with a solution to set vlans using it
1
19,684
13,389,959,492
IssuesEvent
2020-09-02 19:44:59
revangel/nostalgiabot2
https://api.github.com/repos/revangel/nostalgiabot2
opened
Add automatic linting
Technical infrastructure
# Overview It'd be nice to have our code consistently formatted, and more importantly, automatically. Use a combination of flake-8 and [black](https://github.com/psf/black) to implement this # Acceptance Criteria - [ ] A pre-commit hook exists to automatically format code changes
1.0
Add automatic linting - # Overview It'd be nice to have our code consistently formatted, and more importantly, automatically. Use a combination of flake-8 and [black](https://github.com/psf/black) to implement this # Acceptance Criteria - [ ] A pre-commit hook exists to automatically format code changes
non_reli
add automatic linting overview it d be nice to have our code consistently formatted and more importantly automatically use a combination of flake and to implement this acceptance criteria a pre commit hook exists to automatically format code changes
0
588
8,743,752,664
IssuesEvent
2018-12-12 20:05:24
rook/rook
https://api.github.com/repos/rook/rook
closed
Support user-space driver
block perf reliability wontfix
<!-- **Are you in the right place?** 1. For issues or feature requests, please create an issue in this repository. 2. For general technical and non-technical questions, we are happy to help you on our [Rook.io Slack](https://Rook-io.slack.com). 3. Did you already search the existing open issues for anything similar? --> **Is this a bug report or feature request?** * Feature Request **What should the feature do:** I have tried to use `rbd-nbd` feature-rich client to support ceph rbd image. But I noticed it was restricted by existing mechanisms. When using rbd-nbd to map rbd image, it would create a user-space process to deal with network IO. Currently ceph client map rbd image in the agent pod. If I use `rbd-nbd` client in the same way, I found that 1. `rbd-nbd list-mapped` not work. Because the rbd-nbd process have different process id in the container and the host. 2. the rook agent became not stateless. The rbd-nbd process would prevent agent stop/restar. I think any user-space driver like `rbd-nbd` would face similar problems. Because the key reason is that the driver process is running in the container but not the host. I hope we can build an unified architecture to support this type driver. It is more suitable that any driver that use user-space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly. **What is use case behind this feature:** **Environment**: <!-- Specific environment information that helps with the feature request -->
True
Support user-space driver - <!-- **Are you in the right place?** 1. For issues or feature requests, please create an issue in this repository. 2. For general technical and non-technical questions, we are happy to help you on our [Rook.io Slack](https://Rook-io.slack.com). 3. Did you already search the existing open issues for anything similar? --> **Is this a bug report or feature request?** * Feature Request **What should the feature do:** I have tried to use `rbd-nbd` feature-rich client to support ceph rbd image. But I noticed it was restricted by existing mechanisms. When using rbd-nbd to map rbd image, it would create a user-space process to deal with network IO. Currently ceph client map rbd image in the agent pod. If I use `rbd-nbd` client in the same way, I found that 1. `rbd-nbd list-mapped` not work. Because the rbd-nbd process have different process id in the container and the host. 2. the rook agent became not stateless. The rbd-nbd process would prevent agent stop/restar. I think any user-space driver like `rbd-nbd` would face similar problems. Because the key reason is that the driver process is running in the container but not the host. I hope we can build an unified architecture to support this type driver. It is more suitable that any driver that use user-space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly. **What is use case behind this feature:** **Environment**: <!-- Specific environment information that helps with the feature request -->
reli
support user space driver are you in the right place for issues or feature requests please create an issue in this repository for general technical and non technical questions we are happy to help you on our did you already search the existing open issues for anything similar is this a bug report or feature request feature request what should the feature do i have tried to use rbd nbd feature rich client to support ceph rbd image but i noticed it was restricted by existing mechanisms when using rbd nbd to map rbd image it would create a user space process to deal with network io currently ceph client map rbd image in the agent pod if i use rbd nbd client in the same way i found that rbd nbd list mapped not work because the rbd nbd process have different process id in the container and the host the rook agent became not stateless the rbd nbd process would prevent agent stop restar i think any user space driver like rbd nbd would face similar problems because the key reason is that the driver process is running in the container but not the host i hope we can build an unified architecture to support this type driver it is more suitable that any driver that use user space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly what is use case behind this feature environment
1
2,315
24,682,950,719
IssuesEvent
2022-10-18 23:42:26
microsoft/pxt-arcade
https://api.github.com/repos/microsoft/pxt-arcade
closed
"maximum file size in package is ~64k" error causes game crash
bug p2 reliability forum
See: https://forum.makecode.com/t/maximum-file-size-in-package-is-64k/5267/2 I'm asking for more details and will update if we learn more.
True
"maximum file size in package is ~64k" error causes game crash - See: https://forum.makecode.com/t/maximum-file-size-in-package-is-64k/5267/2 I'm asking for more details and will update if we learn more.
reli
maximum file size in package is error causes game crash see i m asking for more details and will update if we learn more
1
2,731
27,298,852,026
IssuesEvent
2023-02-23 23:05:39
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
opened
[FEA] Run CI tests with injected OOMs
feature request ? - Needs Triage reliability
The idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically. A retried test that passes CI (where the output is the same as the CPU) is very valuable and much better at stressing this code out, with the alternative being replicated unit tests that touch a narrower scope. I have prototyped this locally, but can't really PR it yet because I need some hooks from `RmmSpark` and thread association/disassociation that @revans2 is plumbing with the retry framework already. I am thinking about this in the context of code using `withRetry` (https://github.com/NVIDIA/spark-rapids/issues/7256). This gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions. Code outside `withRetry` doesn't have a natural entry points that I can think of, and it wouldn't know how to retry. We need the tests to be very loud about when they are getting an injected OOM. What I have prototyped adds "INJECT_OOM" in the test name, so it should be really easy to figure out that a failure is likely related to OOM handling. ``` ../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[32][INJECT_OOM] ``` vs ``` ../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[33] ```
True
[FEA] Run CI tests with injected OOMs - The idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically. A retried test that passes CI (where the output is the same as the CPU) is very valuable and much better at stressing this code out, with the alternative being replicated unit tests that touch a narrower scope. I have prototyped this locally, but can't really PR it yet because I need some hooks from `RmmSpark` and thread association/disassociation that @revans2 is plumbing with the retry framework already. I am thinking about this in the context of code using `withRetry` (https://github.com/NVIDIA/spark-rapids/issues/7256). This gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions. Code outside `withRetry` doesn't have a natural entry points that I can think of, and it wouldn't know how to retry. We need the tests to be very loud about when they are getting an injected OOM. What I have prototyped adds "INJECT_OOM" in the test name, so it should be really easy to figure out that a failure is likely related to OOM handling. ``` ../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[32][INJECT_OOM] ``` vs ``` ../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[33] ```
reli
run ci tests with injected ooms the idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically a retried test that passes ci where the output is the same as the cpu is very valuable and much better at stressing this code out with the alternative being replicated unit tests that touch a narrower scope i have prototyped this locally but can t really pr it yet because i need some hooks from rmmspark and thread association disassociation that is plumbing with the retry framework already i am thinking about this in the context of code using withretry this gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions code outside withretry doesn t have a natural entry points that i can think of and it wouldn t know how to retry we need the tests to be very loud about when they are getting an injected oom what i have prototyped adds inject oom in the test name so it should be really easy to figure out that a failure is likely related to oom handling src main python hash aggregate test py test hash reduction decimal overflow sum vs src main python hash aggregate test py test hash reduction decimal overflow sum
1
733,113
25,288,537,765
IssuesEvent
2022-11-16 21:34:06
thoth-station/kebechet
https://api.github.com/repos/thoth-station/kebechet
closed
descriptive text is wrong
kind/bug priority/important-soon sig/user-experience
## Bug description The descriptive text of the first bullet point at https://github.com/thoth-station/thamos/pull/1180 is commenting on the wrong object. the text is on a PR, but describes what happens on an issue. ### Steps to Reproduce Steps to reproduce the behavior: 1. Go to https://github.com/thoth-station/thamos/pull/1180 2. see body of PR 4. See error ### Actual behavior won't object used in sentence. ### Expected behavior PR should be the object ### Environment information n/a ### Additional context n/a /priority important-soon /sig user-experience
1.0
descriptive text is wrong - ## Bug description The descriptive text of the first bullet point at https://github.com/thoth-station/thamos/pull/1180 is commenting on the wrong object. the text is on a PR, but describes what happens on an issue. ### Steps to Reproduce Steps to reproduce the behavior: 1. Go to https://github.com/thoth-station/thamos/pull/1180 2. see body of PR 4. See error ### Actual behavior won't object used in sentence. ### Expected behavior PR should be the object ### Environment information n/a ### Additional context n/a /priority important-soon /sig user-experience
non_reli
descriptive text is wrong bug description the descriptive text of the first bullet point at is commenting on the wrong object the text is on a pr but describes what happens on an issue steps to reproduce steps to reproduce the behavior go to see body of pr see error actual behavior won t object used in sentence expected behavior pr should be the object environment information n a additional context n a priority important soon sig user experience
0
1,698
18,836,732,741
IssuesEvent
2021-11-11 02:27:46
zer0Kerbal/OhScrap
https://api.github.com/repos/zer0Kerbal/OhScrap
opened
Oh Scrap!/Scrapyard support?
mod support reliability
From Kerbalism created by [baldamundo](https://github.com/baldamundo): Kerbalism/Kerbalism#493 Wondered if anyone has given this any thought? Oh Scrap (https://forum.kerbalspaceprogram.com/index.php?/topic/160854-17x-oh-scrap-a-scrapyard-based-part-failure-and-reliability-mod-163-22062019/) is an incredibly good part failure mod, which ties the failure to how many times you've built copies of that part, as well as how often that particular instance of the part has been re-used. This gives a huge amount of depth to a career game, encouraging using common hardware across your stable of vehicles, giving an actual reason for test-flights, significantly nerfing/balancing reuseable vehicles, etc. At present, the two mods seem to be perfectly compatible, and actually seem to complement each other quite nicely since the Oh Scrap failures only happen while a vessel is loaded - so generally concentrated around launch failures - while Kerbalism's happen in the background and thus better representing long-term decay. Ultimately the Oh Scrap system is much deeper than Kerbalism's, but Kerbalism has two advantages: 1. Part failures in Kerbalism can occur during background simulation. 2. Failed parts in Kerbalism are displayed in the Kerbalism UI. Is it at all feasible to integrate the two systems better? I'm guessing that either adding background simulation to Oh Scrap or part history tracking to Kerbalism would be very much non-trivial, but at a minimum, is there a way for Oh Scrap to plug into Kerbalism's API and show its failures in the Kerbalism UI?
True
Oh Scrap!/Scrapyard support? - From Kerbalism created by [baldamundo](https://github.com/baldamundo): Kerbalism/Kerbalism#493 Wondered if anyone has given this any thought? Oh Scrap (https://forum.kerbalspaceprogram.com/index.php?/topic/160854-17x-oh-scrap-a-scrapyard-based-part-failure-and-reliability-mod-163-22062019/) is an incredibly good part failure mod, which ties the failure to how many times you've built copies of that part, as well as how often that particular instance of the part has been re-used. This gives a huge amount of depth to a career game, encouraging using common hardware across your stable of vehicles, giving an actual reason for test-flights, significantly nerfing/balancing reuseable vehicles, etc. At present, the two mods seem to be perfectly compatible, and actually seem to complement each other quite nicely since the Oh Scrap failures only happen while a vessel is loaded - so generally concentrated around launch failures - while Kerbalism's happen in the background and thus better representing long-term decay. Ultimately the Oh Scrap system is much deeper than Kerbalism's, but Kerbalism has two advantages: 1. Part failures in Kerbalism can occur during background simulation. 2. Failed parts in Kerbalism are displayed in the Kerbalism UI. Is it at all feasible to integrate the two systems better? I'm guessing that either adding background simulation to Oh Scrap or part history tracking to Kerbalism would be very much non-trivial, but at a minimum, is there a way for Oh Scrap to plug into Kerbalism's API and show its failures in the Kerbalism UI?
reli
oh scrap scrapyard support from kerbalism created by kerbalism kerbalism wondered if anyone has given this any thought oh scrap is an incredibly good part failure mod which ties the failure to how many times you ve built copies of that part as well as how often that particular instance of the part has been re used this gives a huge amount of depth to a career game encouraging using common hardware across your stable of vehicles giving an actual reason for test flights significantly nerfing balancing reuseable vehicles etc at present the two mods seem to be perfectly compatible and actually seem to complement each other quite nicely since the oh scrap failures only happen while a vessel is loaded so generally concentrated around launch failures while kerbalism s happen in the background and thus better representing long term decay ultimately the oh scrap system is much deeper than kerbalism s but kerbalism has two advantages part failures in kerbalism can occur during background simulation failed parts in kerbalism are displayed in the kerbalism ui is it at all feasible to integrate the two systems better i m guessing that either adding background simulation to oh scrap or part history tracking to kerbalism would be very much non trivial but at a minimum is there a way for oh scrap to plug into kerbalism s api and show its failures in the kerbalism ui
1
125,845
17,861,250,044
IssuesEvent
2021-09-06 01:00:06
maorkuriel/JavaVulnerableLab
https://api.github.com/repos/maorkuriel/JavaVulnerableLab
opened
CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar
security vulnerability
## CVE-2015-4852 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: JavaVulnerableLab/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar,JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - hibernate-core-4.0.1.Final.jar (Root Library) - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product. <p>Publish Date: 2015-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p> <p>Release Date: 2015-11-18</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.hibernate:hibernate-core:4.0.1.Final;commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-collections:commons-collections:3.2.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-4852","vulnerabilityDetails":"The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
True
CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-4852 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: JavaVulnerableLab/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar,JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - hibernate-core-4.0.1.Final.jar (Root Library) - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product. <p>Publish Date: 2015-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p> <p>Release Date: 2015-11-18</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.hibernate:hibernate-core:4.0.1.Final;commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-collections:commons-collections:3.2.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-4852","vulnerabilityDetails":"The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
non_reli
cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file javavulnerablelab pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar javavulnerablelab web inf lib commons collections jar dependency hierarchy hibernate core final jar root library x commons collections jar vulnerable library found in base branch master vulnerability details the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org hibernate hibernate core final commons collections commons collections isminimumfixversionavailable true minimumfixversion commons collections commons collections basebranches vulnerabilityidentifier cve vulnerabilitydetails the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product vulnerabilityurl
0
2,761
27,557,522,324
IssuesEvent
2023-03-07 19:08:28
rapidsai/cudf
https://api.github.com/repos/rapidsai/cudf
opened
[FEA][JNI] Leverage cub's multi-buffer copy algorithm in JNI bindings
feature request ? - Needs Triage Performance cuDF (Java) helps: Spark Reliability
We would like to use https://github.com/NVIDIA/cub/issues/297 for our spilling logic https://github.com/NVIDIA/spark-rapids/issues/7672 in the spark-rapids plugin. We currently contiguous_split every single buffer that _could be_ spilled in the future, because at the time of spill the last thing we want to do is double our memory usage right before copying to host. The multi-buffer work should allow us to bounce-buffer out of GPU memory at spill time, without having to contiguous_split for spillable buffers.
True
[FEA][JNI] Leverage cub's multi-buffer copy algorithm in JNI bindings - We would like to use https://github.com/NVIDIA/cub/issues/297 for our spilling logic https://github.com/NVIDIA/spark-rapids/issues/7672 in the spark-rapids plugin. We currently contiguous_split every single buffer that _could be_ spilled in the future, because at the time of spill the last thing we want to do is double our memory usage right before copying to host. The multi-buffer work should allow us to bounce-buffer out of GPU memory at spill time, without having to contiguous_split for spillable buffers.
reli
leverage cub s multi buffer copy algorithm in jni bindings we would like to use for our spilling logic in the spark rapids plugin we currently contiguous split every single buffer that could be spilled in the future because at the time of spill the last thing we want to do is double our memory usage right before copying to host the multi buffer work should allow us to bounce buffer out of gpu memory at spill time without having to contiguous split for spillable buffers
1
50,032
7,551,601,394
IssuesEvent
2018-04-18 20:40:23
kubernetes/kubeadm
https://api.github.com/repos/kubernetes/kubeadm
closed
Please document that self-hosted feature does not survive a reboot (v 1.9.3)
active kind/documentation kind/feature priority/important-soon
<!-- Thanks for filing an issue! Before hitting the button, please answer these questions.--> ## Is this a BUG REPORT or FEATURE REQUEST? FEATURE REQUEST ## Versions **kubeadm version** (use `kubeadm version`): ``` core@localhost ~ $ kubeadm version kubeadm version: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} ``` **Environment**: - **Kubernetes version** (use `kubectl version`): ``` core@localhost ~ $ kubectl version Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T12:22:21Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} ``` - **Cloud provider or hardware configuration**: Dell R720 Bare Metal Server - **OS** (e.g. from /etc/os-release): ``` core@localhost ~ $ cat /etc/os-release NAME="Container Linux by CoreOS" ID=coreos VERSION=1632.3.0 VERSION_ID=1632.3.0 BUILD_ID=2018-02-14-0338 PRETTY_NAME="Container Linux by CoreOS 1632.3.0 (Ladybug)" ANSI_COLOR="38;5;75" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://issues.coreos.com" COREOS_BOARD="amd64-usr" ``` - **Kernel** (e.g. `uname -a`): ``` core@localhost ~ $ uname -a Linux localhost 4.14.19-coreos #1 SMP Wed Feb 14 03:18:05 UTC 2018 x86_64 Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz GenuineIntel GNU/Linux ``` - **Others**: ## What happened? Please document that the self-hosted feature does not work when you reboot the server without additional configuration. Kubernetes cluster does not restart after reboot ## What you expected to happen? Expected it to restart. ## How to reproduce it (as minimally and precisely as possible)? ``` sudo kubeadm init --feature-gates=SelfHosting=true --pod-network-cidr=10.244.0.0/16 sudo systemctl enable kubelet.service sudo shutdown -r now ``` ## Anything else we need to know? A workaround is described here: https://github.com/xetys/k8s-self-hosted-recovery It took me a long time to find it because I didn't understand what was going wrong so it was hard to search. Once I understood that the API service wasn't starting and it was a catch-22 problem due to the self-hosted feature, then it was pretty each to search and find the workaround.
1.0
Please document that self-hosted feature does not survive a reboot (v 1.9.3) - <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.--> ## Is this a BUG REPORT or FEATURE REQUEST? FEATURE REQUEST ## Versions **kubeadm version** (use `kubeadm version`): ``` core@localhost ~ $ kubeadm version kubeadm version: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} ``` **Environment**: - **Kubernetes version** (use `kubectl version`): ``` core@localhost ~ $ kubectl version Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T12:22:21Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} ``` - **Cloud provider or hardware configuration**: Dell R720 Bare Metal Server - **OS** (e.g. from /etc/os-release): ``` core@localhost ~ $ cat /etc/os-release NAME="Container Linux by CoreOS" ID=coreos VERSION=1632.3.0 VERSION_ID=1632.3.0 BUILD_ID=2018-02-14-0338 PRETTY_NAME="Container Linux by CoreOS 1632.3.0 (Ladybug)" ANSI_COLOR="38;5;75" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://issues.coreos.com" COREOS_BOARD="amd64-usr" ``` - **Kernel** (e.g. `uname -a`): ``` core@localhost ~ $ uname -a Linux localhost 4.14.19-coreos #1 SMP Wed Feb 14 03:18:05 UTC 2018 x86_64 Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz GenuineIntel GNU/Linux ``` - **Others**: ## What happened? Please document that the self-hosted feature does not work when you reboot the server without additional configuration. Kubernetes cluster does not restart after reboot ## What you expected to happen? Expected it to restart. ## How to reproduce it (as minimally and precisely as possible)? ``` sudo kubeadm init --feature-gates=SelfHosting=true --pod-network-cidr=10.244.0.0/16 sudo systemctl enable kubelet.service sudo shutdown -r now ``` ## Anything else we need to know? A workaround is described here: https://github.com/xetys/k8s-self-hosted-recovery It took me a long time to find it because I didn't understand what was going wrong so it was hard to search. Once I understood that the API service wasn't starting and it was a catch-22 problem due to the self-hosted feature, then it was pretty each to search and find the workaround.
non_reli
please document that self hosted feature does not survive a reboot v is this a bug report or feature request feature request versions kubeadm version use kubeadm version core localhost kubeadm version kubeadm version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux environment kubernetes version use kubectl version core localhost kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux cloud provider or hardware configuration dell bare metal server os e g from etc os release core localhost cat etc os release name container linux by coreos id coreos version version id build id pretty name container linux by coreos ladybug ansi color home url bug report url coreos board usr kernel e g uname a core localhost uname a linux localhost coreos smp wed feb utc intel r xeon r cpu genuineintel gnu linux others what happened please document that the self hosted feature does not work when you reboot the server without additional configuration kubernetes cluster does not restart after reboot what you expected to happen expected it to restart how to reproduce it as minimally and precisely as possible sudo kubeadm init feature gates selfhosting true pod network cidr sudo systemctl enable kubelet service sudo shutdown r now anything else we need to know a workaround is described here it took me a long time to find it because i didn t understand what was going wrong so it was hard to search once i understood that the api service wasn t starting and it was a catch problem due to the self hosted feature then it was pretty each to search and find the workaround
0
29,629
24,122,163,495
IssuesEvent
2022-09-20 19:46:36
GCTC-NTGC/gc-digital-talent
https://api.github.com/repos/GCTC-NTGC/gc-digital-talent
closed
Not possible to remove a feature flag on a single vertical
infrastructure
When removing a feature flag it would be good to be able to do it progressively: dev, UAT, and finally prod. Right now the setting is defined in appsvc-VERTICAL-talentcloud-template.json which is the shared file. So it isn't possible, I think.
1.0
Not possible to remove a feature flag on a single vertical - When removing a feature flag it would be good to be able to do it progressively: dev, UAT, and finally prod. Right now the setting is defined in appsvc-VERTICAL-talentcloud-template.json which is the shared file. So it isn't possible, I think.
non_reli
not possible to remove a feature flag on a single vertical when removing a feature flag it would be good to be able to do it progressively dev uat and finally prod right now the setting is defined in appsvc vertical talentcloud template json which is the shared file so it isn t possible i think
0
1,932
21,897,270,176
IssuesEvent
2022-05-20 09:50:28
jina-ai/jina
https://api.github.com/repos/jina-ai/jina
closed
Smarter retry mechanism for Executors with replicas
epic/reliability
Currently, the round robin scheduling between replicas does not apply when retrying to send a request, but only when a new request comes in. This means that if a replica is down, the gateway stubbornly tries to connect to the this same replica, ultimately failing the request despite other replicas being available. This should be changed s.t. round robing scheduling is applied also to connection retries.
True
Smarter retry mechanism for Executors with replicas - Currently, the round robin scheduling between replicas does not apply when retrying to send a request, but only when a new request comes in. This means that if a replica is down, the gateway stubbornly tries to connect to the this same replica, ultimately failing the request despite other replicas being available. This should be changed s.t. round robing scheduling is applied also to connection retries.
reli
smarter retry mechanism for executors with replicas currently the round robin scheduling between replicas does not apply when retrying to send a request but only when a new request comes in this means that if a replica is down the gateway stubbornly tries to connect to the this same replica ultimately failing the request despite other replicas being available this should be changed s t round robing scheduling is applied also to connection retries
1
757,487
26,514,714,957
IssuesEvent
2023-01-18 19:49:56
ramp4-pcar4/storylines-editor
https://api.github.com/repos/ramp4-pcar4/storylines-editor
closed
Editor slides do not switch properly when switching between panels of the same type
Bug Priority: High Needs: estimate
If you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type, the contents of the page are not updated properly. This is an issue we had with dynamic panels in the main Storylines app too, has something to do with Vue not re-rendering everything. Steps to Reproduce: 1. Click on the `Oil sands extraction` slide and switch to the right panel. 2. Click on the `In-situ extraction` slide. 3. Notice that the image displayed is still for `Oil Sands Extraction`. *Demo* ![](https://i.imgur.com/ehb4380.gif)
1.0
Editor slides do not switch properly when switching between panels of the same type - If you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type, the contents of the page are not updated properly. This is an issue we had with dynamic panels in the main Storylines app too, has something to do with Vue not re-rendering everything. Steps to Reproduce: 1. Click on the `Oil sands extraction` slide and switch to the right panel. 2. Click on the `In-situ extraction` slide. 3. Notice that the image displayed is still for `Oil Sands Extraction`. *Demo* ![](https://i.imgur.com/ehb4380.gif)
non_reli
editor slides do not switch properly when switching between panels of the same type if you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type the contents of the page are not updated properly this is an issue we had with dynamic panels in the main storylines app too has something to do with vue not re rendering everything steps to reproduce click on the oil sands extraction slide and switch to the right panel click on the in situ extraction slide notice that the image displayed is still for oil sands extraction demo
0
214,874
24,121,154,245
IssuesEvent
2022-09-20 18:50:25
Azure/AKS
https://api.github.com/repos/Azure/AKS
closed
Add Support for a KMS provider for Encrypting Secrets
security feature-request resolution/shipped
According to https://github.com/Azure/kubernetes-kms: > We have added this feature to aks-engine so that you do not have to worry about any of the manual steps to set this up. Does this mean Azure AKS encrypts secrets by default? I don't see a Key Vault being setup in the AKS resource group, so where are the keys stored? The above link also says: > Currently KMS plugin for Kubernetes does not support KMS key rotation scenarios. So keys are never rotated and hard coded from the start?
True
Add Support for a KMS provider for Encrypting Secrets - According to https://github.com/Azure/kubernetes-kms: > We have added this feature to aks-engine so that you do not have to worry about any of the manual steps to set this up. Does this mean Azure AKS encrypts secrets by default? I don't see a Key Vault being setup in the AKS resource group, so where are the keys stored? The above link also says: > Currently KMS plugin for Kubernetes does not support KMS key rotation scenarios. So keys are never rotated and hard coded from the start?
non_reli
add support for a kms provider for encrypting secrets according to we have added this feature to aks engine so that you do not have to worry about any of the manual steps to set this up does this mean azure aks encrypts secrets by default i don t see a key vault being setup in the aks resource group so where are the keys stored the above link also says currently kms plugin for kubernetes does not support kms key rotation scenarios so keys are never rotated and hard coded from the start
0
137,202
20,101,659,101
IssuesEvent
2022-02-07 05:25:24
Facepunch/sbox-issues
https://api.github.com/repos/Facepunch/sbox-issues
opened
Add WithSize, WithFormat, etc to TextureCubeBuilder
api design
**What it is** Currently, TextureCubeBuilder only has the methods WithUAVBinding and Finish. Texture.CreateCube doesn't take any arguments either. As far as I can tell, there's no way to specify the width/height of the cube texture, or the format of the data passed in. As a result, it is currently impossible to create cube textures (spits out an error about the texture having 0x0 size). **What it should be** TextureCubeBuilder should have the relevant `WithSize(int width, int height)`, `WithFormat(ImageFormat format)`, `WithMips(int mips)`, etc methods, similar to Texture2DBuilder. Ideally this functionality would also be added to TextureArrayBuilder.
1.0
Add WithSize, WithFormat, etc to TextureCubeBuilder - **What it is** Currently, TextureCubeBuilder only has the methods WithUAVBinding and Finish. Texture.CreateCube doesn't take any arguments either. As far as I can tell, there's no way to specify the width/height of the cube texture, or the format of the data passed in. As a result, it is currently impossible to create cube textures (spits out an error about the texture having 0x0 size). **What it should be** TextureCubeBuilder should have the relevant `WithSize(int width, int height)`, `WithFormat(ImageFormat format)`, `WithMips(int mips)`, etc methods, similar to Texture2DBuilder. Ideally this functionality would also be added to TextureArrayBuilder.
non_reli
add withsize withformat etc to texturecubebuilder what it is currently texturecubebuilder only has the methods withuavbinding and finish texture createcube doesn t take any arguments either as far as i can tell there s no way to specify the width height of the cube texture or the format of the data passed in as a result it is currently impossible to create cube textures spits out an error about the texture having size what it should be texturecubebuilder should have the relevant withsize int width int height withformat imageformat format withmips int mips etc methods similar to ideally this functionality would also be added to texturearraybuilder
0
841
10,789,348,390
IssuesEvent
2019-11-05 11:43:06
sohaibaslam/learning_site
https://api.github.com/repos/sohaibaslam/learning_site
opened
Broken Crawler: puma
crawler broken/unreliable
name: puma-at name: puma-be name: puma-ch name: puma-cz name: puma-de name: puma-dk name: puma-es name: puma-fi name: puma-fr name: puma-gr name: puma-hu name: puma-ie name: puma-it name: puma-lu name: puma-nl name: puma-pl name: puma-pt name: puma-se mixin_count: 18
True
Broken Crawler: puma - name: puma-at name: puma-be name: puma-ch name: puma-cz name: puma-de name: puma-dk name: puma-es name: puma-fi name: puma-fr name: puma-gr name: puma-hu name: puma-ie name: puma-it name: puma-lu name: puma-nl name: puma-pl name: puma-pt name: puma-se mixin_count: 18
reli
broken crawler puma name puma at name puma be name puma ch name puma cz name puma de name puma dk name puma es name puma fi name puma fr name puma gr name puma hu name puma ie name puma it name puma lu name puma nl name puma pl name puma pt name puma se mixin count
1
165,437
14,002,060,981
IssuesEvent
2020-10-28 14:22:29
AY2021S1-CS2103T-T09-1/tp
https://api.github.com/repos/AY2021S1-CS2103T-T09-1/tp
closed
UG: Add Tags command + update add, edit, find commands
documentation priority.High type.Task
- Add Tags command - Update add command - Update edit command - Update find command
1.0
UG: Add Tags command + update add, edit, find commands - - Add Tags command - Update add command - Update edit command - Update find command
non_reli
ug add tags command update add edit find commands add tags command update add command update edit command update find command
0
259,874
8,200,923,745
IssuesEvent
2018-09-01 11:14:52
TuranicTeam/Altay
https://api.github.com/repos/TuranicTeam/Altay
closed
Banner's
Category: Core Priority: Low Type: Cosmetic
#### Issue description It would be necessary to operate banner crafts! [ ![image](https://user-images.githubusercontent.com/28710769/41804503-bd0fca02-7697-11e8-8afc-179e49586965.png) ](url) #### Steps to reproduce the issue <!--- Help us find the problem by adding steps to reproduce the issue --> - #### OS and versions <!--- Valid version must contain build number or git hash, "latest" is invalid. If the version is invalid, the issue will be CLOSED Game versions include Bedrock Engine, Windows 10 Edition version --> * Altay: 3.0.7+dev * PHP: 7.2.8 * Server OS: Windows 10 * Game version: 1.5.2 Windows 10 Edition #### Crashdump, backtrace, memory dumps, plugins list or other files <!--- Paste in the below block --> ``` - ```
1.0
Banner's - #### Issue description It would be necessary to operate banner crafts! [ ![image](https://user-images.githubusercontent.com/28710769/41804503-bd0fca02-7697-11e8-8afc-179e49586965.png) ](url) #### Steps to reproduce the issue <!--- Help us find the problem by adding steps to reproduce the issue --> - #### OS and versions <!--- Valid version must contain build number or git hash, "latest" is invalid. If the version is invalid, the issue will be CLOSED Game versions include Bedrock Engine, Windows 10 Edition version --> * Altay: 3.0.7+dev * PHP: 7.2.8 * Server OS: Windows 10 * Game version: 1.5.2 Windows 10 Edition #### Crashdump, backtrace, memory dumps, plugins list or other files <!--- Paste in the below block --> ``` - ```
non_reli
banner s issue description it would be necessary to operate banner crafts url steps to reproduce the issue os and versions valid version must contain build number or git hash latest is invalid if the version is invalid the issue will be closed game versions include bedrock engine windows edition version altay dev php server os windows game version windows edition crashdump backtrace memory dumps plugins list or other files
0
165,756
12,879,870,884
IssuesEvent
2020-07-12 01:26:52
osquery/osquery
https://api.github.com/repos/osquery/osquery
closed
Create tests for the table `sharing_preferences`
good-first-issue macOS test
## Create tests for the table `sharing_preferences` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/darwin/sharing_preferences.table` Source files: - `osquery/tables/system/darwin/sharing_preferences.cpp` Table generating function: `genSharingPreferences()`
1.0
Create tests for the table `sharing_preferences` - ## Create tests for the table `sharing_preferences` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/darwin/sharing_preferences.table` Source files: - `osquery/tables/system/darwin/sharing_preferences.cpp` Table generating function: `genSharingPreferences()`
non_reli
create tests for the table sharing preferences create tests for the table sharing preferences create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs darwin sharing preferences table source files osquery tables system darwin sharing preferences cpp table generating function gensharingpreferences
0
154
4,628,109,909
IssuesEvent
2016-09-28 02:20:03
UofSSpaceDesignTeam/rover-hardware
https://api.github.com/repos/UofSSpaceDesignTeam/rover-hardware
closed
Replace bad battery cell
Mandatory Power Reliability
One of the battery cells started expanding. It has been removed, and will be taken away to be (hopefully) disposed of properly. A new one will need to be installed to replace it. A new case will need to be designed to hold the cells under the right amount of pressure (job for the mechanical team?).
True
Replace bad battery cell - One of the battery cells started expanding. It has been removed, and will be taken away to be (hopefully) disposed of properly. A new one will need to be installed to replace it. A new case will need to be designed to hold the cells under the right amount of pressure (job for the mechanical team?).
reli
replace bad battery cell one of the battery cells started expanding it has been removed and will be taken away to be hopefully disposed of properly a new one will need to be installed to replace it a new case will need to be designed to hold the cells under the right amount of pressure job for the mechanical team
1
471
7,734,135,141
IssuesEvent
2018-05-26 20:27:07
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Cloud provider api issues can prevent kubelet from reporting node status
area/reliability kind/bug lifecycle/rotten sig/node
There have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes. If the cloud provider api is unresponsive, it could prevent the node from reporting node/pod status to the apiserver. If this happens for too long, the node will be marked NotReady even though the kubelet is actually running without any issues. I don't think we want to prevent a node or pod update because the cloud provider api has become unresponsive. We should try to ensure a node only gets marked NotReady if it is truly incapable of operating. xref: #43516 #46602 #46340 @dchen1107 @kubernetes/sig-node-bugs
True
Cloud provider api issues can prevent kubelet from reporting node status - There have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes. If the cloud provider api is unresponsive, it could prevent the node from reporting node/pod status to the apiserver. If this happens for too long, the node will be marked NotReady even though the kubelet is actually running without any issues. I don't think we want to prevent a node or pod update because the cloud provider api has become unresponsive. We should try to ensure a node only gets marked NotReady if it is truly incapable of operating. xref: #43516 #46602 #46340 @dchen1107 @kubernetes/sig-node-bugs
reli
cloud provider api issues can prevent kubelet from reporting node status there have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes if the cloud provider api is unresponsive it could prevent the node from reporting node pod status to the apiserver if this happens for too long the node will be marked notready even though the kubelet is actually running without any issues i don t think we want to prevent a node or pod update because the cloud provider api has become unresponsive we should try to ensure a node only gets marked notready if it is truly incapable of operating xref kubernetes sig node bugs
1
2,080
23,229,350,057
IssuesEvent
2022-08-03 05:50:57
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Infinite recursion during resource lookup within System.Private.CoreLib
area-Meta tenet-reliability untriaged needs-author-action
### Description ``` CoreCLR Version: 6.0.622.26707 .NET Version: 6.0.6 Description: The application requested process termination through System.Environment.FailFast(string message). Message: Infinite recursion during resource lookup within System.Private.CoreLib. This may be a bug in System.Private.CoreLib, or potentially in certain extensibility points such as assembly resolve events or CultureInfo names. Resource name: Arg_NullReferenceException Stack: at System.Environment.FailFast(System.String) at System.SR.InternalGetResourceString(System.String) at System.SR.GetResourceString(System.String) at System.NullReferenceException..ctor() at System.Globalization.CultureInfo.get_CurrentUICulture() at System.Resources.ResourceManager.GetString(System.String, System.Globalization.CultureInfo) at System.SR.InternalGetResourceString(System.String) at System.SR.GetResourceString(System.String) at System.NullReferenceException..ctor() at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext(System.Threading.Thread) at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext() at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.InvokeContinuation(System.Action`1<System.Object>, System.Object, Boolean, Boolean) at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.OnCompleted(System.Net.Sockets.SocketAsyncEventArgs) at System.Threading._IOCompletionCallback.PerformIOCompletionCallback(UInt32, UInt32, System.Threading.NativeOverlapped*) ``` ### Configuration <!-- * Which version of .NET is the code running on? * What OS and version, and for Linux, what distro? * What is the architecture (x64, x86, ARM, ARM64)? * Do you know whether it is specific to that configuration? * If you're using Blazor, which web browser(s) do you see this issue in? --> ### Regression? <!-- * Did this work in a previous build or release of .NET, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK. --> ### Other information <!-- * Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches). * If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of. * Do you know of any workarounds? -->
True
Infinite recursion during resource lookup within System.Private.CoreLib - ### Description ``` CoreCLR Version: 6.0.622.26707 .NET Version: 6.0.6 Description: The application requested process termination through System.Environment.FailFast(string message). Message: Infinite recursion during resource lookup within System.Private.CoreLib. This may be a bug in System.Private.CoreLib, or potentially in certain extensibility points such as assembly resolve events or CultureInfo names. Resource name: Arg_NullReferenceException Stack: at System.Environment.FailFast(System.String) at System.SR.InternalGetResourceString(System.String) at System.SR.GetResourceString(System.String) at System.NullReferenceException..ctor() at System.Globalization.CultureInfo.get_CurrentUICulture() at System.Resources.ResourceManager.GetString(System.String, System.Globalization.CultureInfo) at System.SR.InternalGetResourceString(System.String) at System.SR.GetResourceString(System.String) at System.NullReferenceException..ctor() at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext(System.Threading.Thread) at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext() at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.InvokeContinuation(System.Action`1<System.Object>, System.Object, Boolean, Boolean) at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.OnCompleted(System.Net.Sockets.SocketAsyncEventArgs) at System.Threading._IOCompletionCallback.PerformIOCompletionCallback(UInt32, UInt32, System.Threading.NativeOverlapped*) ``` ### Configuration <!-- * Which version of .NET is the code running on? * What OS and version, and for Linux, what distro? * What is the architecture (x64, x86, ARM, ARM64)? * Do you know whether it is specific to that configuration? * If you're using Blazor, which web browser(s) do you see this issue in? --> ### Regression? <!-- * Did this work in a previous build or release of .NET, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK. --> ### Other information <!-- * Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches). * If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of. * Do you know of any workarounds? -->
reli
infinite recursion during resource lookup within system private corelib description coreclr version net version description the application requested process termination through system environment failfast string message message infinite recursion during resource lookup within system private corelib this may be a bug in system private corelib or potentially in certain extensibility points such as assembly resolve events or cultureinfo names resource name arg nullreferenceexception stack at system environment failfast system string at system sr internalgetresourcestring system string at system sr getresourcestring system string at system nullreferenceexception ctor at system globalization cultureinfo get currentuiculture at system resources resourcemanager getstring system string system globalization cultureinfo at system sr internalgetresourcestring system string at system sr getresourcestring system string at system nullreferenceexception ctor at system threading executioncontext runinternal system threading executioncontext system threading contextcallback system object at system runtime compilerservices asynctaskmethodbuilder asyncstatemachinebox system net security version culture neutral publickeytoken movenext system threading thread at system runtime compilerservices asynctaskmethodbuilder asyncstatemachinebox system net security version culture neutral publickeytoken movenext at system net sockets socket awaitablesocketasynceventargs invokecontinuation system action system object boolean boolean at system net sockets socket awaitablesocketasynceventargs oncompleted system net sockets socketasynceventargs at system threading iocompletioncallback performiocompletioncallback system threading nativeoverlapped configuration which version of net is the code running on what os and version and for linux what distro what is the architecture arm do you know whether it is specific to that configuration if you re using blazor which web browser s do you see this issue in regression did this work in a previous build or release of net or from net framework if you can try a previous release or build to find out that can help us narrow down the problem if you don t know that s ok other information please include any relevant stack traces or error messages if possible please include text as text rather than images so it shows up in searches if you have an idea where the problem might lie let us know that here please include any pointers to code relevant changes or related issues you know of do you know of any workarounds
1
273,459
8,530,913,133
IssuesEvent
2018-11-04 05:16:26
CS2113-AY1819S1-W13-2/main
https://api.github.com/repos/CS2113-AY1819S1-W13-2/main
closed
Wrong prefix format used should reflect invalid command format message instead.
priority.high severity.High type.bug
**Describe the bug** Wrong prefix format used should reflect invalid command format message instead. **To Reproduce** Steps to reproduce the behavior: put additional prefix at the end of the full add command as showed in picture. **Expected behavior** A clear and concise description of what you expected to happen. An invalid command format error message should be prompt instead of an internal error occured. This might confuse end user. **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/35736017/47905988-40af0100-dec3-11e8-9056-9a2c9e2e34bd.png) **Additional context** Add any other context about the problem here. <hr> **Reported by:** @ryanchen2018 **Severity:** Not Specified <sub>[original: nusCS2113-AY1819S1/pe-1#154]</sub>
1.0
Wrong prefix format used should reflect invalid command format message instead. - **Describe the bug** Wrong prefix format used should reflect invalid command format message instead. **To Reproduce** Steps to reproduce the behavior: put additional prefix at the end of the full add command as showed in picture. **Expected behavior** A clear and concise description of what you expected to happen. An invalid command format error message should be prompt instead of an internal error occured. This might confuse end user. **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/35736017/47905988-40af0100-dec3-11e8-9056-9a2c9e2e34bd.png) **Additional context** Add any other context about the problem here. <hr> **Reported by:** @ryanchen2018 **Severity:** Not Specified <sub>[original: nusCS2113-AY1819S1/pe-1#154]</sub>
non_reli
wrong prefix format used should reflect invalid command format message instead describe the bug wrong prefix format used should reflect invalid command format message instead to reproduce steps to reproduce the behavior put additional prefix at the end of the full add command as showed in picture expected behavior a clear and concise description of what you expected to happen an invalid command format error message should be prompt instead of an internal error occured this might confuse end user screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reported by severity not specified
0
281
6,001,208,880
IssuesEvent
2017-06-05 08:28:24
datacite/datacite
https://api.github.com/repos/datacite/datacite
opened
Incident report have service outage
data center member reliability
As a data center manager (or member), I want a complete report after every incident so I can explain what happened to my users.
True
Incident report have service outage - As a data center manager (or member), I want a complete report after every incident so I can explain what happened to my users.
reli
incident report have service outage as a data center manager or member i want a complete report after every incident so i can explain what happened to my users
1
792,040
27,943,705,604
IssuesEvent
2023-03-24 00:03:35
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
[MS] Create monitoring dashboard for MSPileup microservice
New Feature High Priority Stakeholders MSPileup QPrio: High
**Impact of the new feature** New WMCore microservice **Is your feature request related to a problem? Please describe.** Creation of a new microservice to make pileup management more efficient and easier. Subtask of https://github.com/dmwm/WMCore/issues/9779 **Describe the solution you'd like** This is a joint ticket between the WMCore and the CMS Monitoring team, such that we can define and provide a monitoring dashboard for the MSPileup. So, relevant JIRA ticket might have to be created to get the usual process monitoring, as we have for the other WMCore microservices. **Describe alternatives you've considered** We can eventually consider service specific metrics as well, but I'd call that to be an future enhancement and we can properly define that in the future. **Additional context** None
1.0
[MS] Create monitoring dashboard for MSPileup microservice - **Impact of the new feature** New WMCore microservice **Is your feature request related to a problem? Please describe.** Creation of a new microservice to make pileup management more efficient and easier. Subtask of https://github.com/dmwm/WMCore/issues/9779 **Describe the solution you'd like** This is a joint ticket between the WMCore and the CMS Monitoring team, such that we can define and provide a monitoring dashboard for the MSPileup. So, relevant JIRA ticket might have to be created to get the usual process monitoring, as we have for the other WMCore microservices. **Describe alternatives you've considered** We can eventually consider service specific metrics as well, but I'd call that to be an future enhancement and we can properly define that in the future. **Additional context** None
non_reli
create monitoring dashboard for mspileup microservice impact of the new feature new wmcore microservice is your feature request related to a problem please describe creation of a new microservice to make pileup management more efficient and easier subtask of describe the solution you d like this is a joint ticket between the wmcore and the cms monitoring team such that we can define and provide a monitoring dashboard for the mspileup so relevant jira ticket might have to be created to get the usual process monitoring as we have for the other wmcore microservices describe alternatives you ve considered we can eventually consider service specific metrics as well but i d call that to be an future enhancement and we can properly define that in the future additional context none
0
336,273
24,491,492,483
IssuesEvent
2022-10-10 02:46:55
vamsitadikonda/chat-defender-bot
https://api.github.com/repos/vamsitadikonda/chat-defender-bot
closed
Doc generation issues
documentation
Documentation generation is not syncing with GitHub pages. Please update the hook
1.0
Doc generation issues - Documentation generation is not syncing with GitHub pages. Please update the hook
non_reli
doc generation issues documentation generation is not syncing with github pages please update the hook
0
2,840
28,101,638,438
IssuesEvent
2023-03-30 20:00:47
supaglue-labs/supaglue
https://api.github.com/repos/supaglue-labs/supaglue
closed
Make sync history log ids idempotent
reliability
Right now, when we log a new SyncHistory event, we generate an id on the fly (or rather, prisma/DB does). This means that on retry, we create a new set of history log ids. Instead of doing this, we should not be writing another set of SyncHistory events. We can do this by using deterministic UUIDs. We can do this by: 1. using temporal's deterministic UUID generator 2. Generate a "fake UUID" like "sync-id:iteration-number` 3. Generate UUIDv5 with namespace using iteration number
True
Make sync history log ids idempotent - Right now, when we log a new SyncHistory event, we generate an id on the fly (or rather, prisma/DB does). This means that on retry, we create a new set of history log ids. Instead of doing this, we should not be writing another set of SyncHistory events. We can do this by using deterministic UUIDs. We can do this by: 1. using temporal's deterministic UUID generator 2. Generate a "fake UUID" like "sync-id:iteration-number` 3. Generate UUIDv5 with namespace using iteration number
reli
make sync history log ids idempotent right now when we log a new synchistory event we generate an id on the fly or rather prisma db does this means that on retry we create a new set of history log ids instead of doing this we should not be writing another set of synchistory events we can do this by using deterministic uuids we can do this by using temporal s deterministic uuid generator generate a fake uuid like sync id iteration number generate with namespace using iteration number
1
6,181
13,886,002,489
IssuesEvent
2020-10-18 22:34:42
bSchnepp/Feral
https://api.github.com/repos/bSchnepp/Feral
closed
[REGRESSION] - RS232 driver is no longer functioning
architecture feature high-priority
Since the CMake migration (https://github.com/bSchnepp/Feral/commit/79a9c3a2631900b422757ce159204135c7c2842b), it appears that the serial driver no longer functions, and COM1 is empty when the kernel finished booting.
1.0
[REGRESSION] - RS232 driver is no longer functioning - Since the CMake migration (https://github.com/bSchnepp/Feral/commit/79a9c3a2631900b422757ce159204135c7c2842b), it appears that the serial driver no longer functions, and COM1 is empty when the kernel finished booting.
non_reli
driver is no longer functioning since the cmake migration it appears that the serial driver no longer functions and is empty when the kernel finished booting
0
448
7,573,968,079
IssuesEvent
2018-04-23 19:27:01
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Reduce allocation requirements in WriteObjectWorker
4 - In Review Area-Compilers Bug Investigation Required Tenet-Reliability
The dictionaries used by `WriteObjectWorker` can become large, and occasionally fail when the IDE is under memory pressure. We should identify sets of objects which cannot be multiply-referenced within the object getting serialized, and avoid adding these objects to the map. https://github.com/dotnet/roslyn/blob/878ffad23b8b06cb229c9ab31eada7634a473508/src/Compilers/Core/Portable/Serialization/ObjectWriter.cs#L805 :link: Watson (internal issue for this): https://devdiv.visualstudio.com/DevDiv/_workitems/edit/485537
True
Reduce allocation requirements in WriteObjectWorker - The dictionaries used by `WriteObjectWorker` can become large, and occasionally fail when the IDE is under memory pressure. We should identify sets of objects which cannot be multiply-referenced within the object getting serialized, and avoid adding these objects to the map. https://github.com/dotnet/roslyn/blob/878ffad23b8b06cb229c9ab31eada7634a473508/src/Compilers/Core/Portable/Serialization/ObjectWriter.cs#L805 :link: Watson (internal issue for this): https://devdiv.visualstudio.com/DevDiv/_workitems/edit/485537
reli
reduce allocation requirements in writeobjectworker the dictionaries used by writeobjectworker can become large and occasionally fail when the ide is under memory pressure we should identify sets of objects which cannot be multiply referenced within the object getting serialized and avoid adding these objects to the map link watson internal issue for this
1
622,429
19,635,153,302
IssuesEvent
2022-01-08 05:57:37
Treetrain1/Survival-Remastered
https://api.github.com/repos/Treetrain1/Survival-Remastered
closed
On exiting the end portal, you will teleport into the void and die
bug high priority
When exiting the end through the portal using beta 5, you will probably fall through the world and die
1.0
On exiting the end portal, you will teleport into the void and die - When exiting the end through the portal using beta 5, you will probably fall through the world and die
non_reli
on exiting the end portal you will teleport into the void and die when exiting the end through the portal using beta you will probably fall through the world and die
0
708
9,980,573,596
IssuesEvent
2019-07-10 04:11:33
dotnet/project-system
https://api.github.com/repos/dotnet/project-system
closed
Renamed an item from an import crashes VS
Bug Feature-Project-File-Simplification Project-System-CPS Tenet-Reliability Urgency-Soon
Renamed an item from an import ``` Application: devenv.exe Framework Version: v4.0.30319 Description: The process was terminated due to an unhandled exception. Exception Info: System.NullReferenceException at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<IsCompilationCaseSensitiveAsync>d__15.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task) at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<RenameAsync>d__13.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.ValidateEnd(System.Threading.Tasks.Task) at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<OnWorkspaceChangedAsync>d__12.MoveNext() at System.Runtime.CompilerServices.AsyncMethodBuilderCore+<>c.<ThrowAsync>b__6_0(System.Object) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate) at System.Windows.Threading.DispatcherOperation.InvokeImpl() at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(System.Object) at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(System.Object) at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean) at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean) at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at MS.Internal.CulturePreservingExecutionContext.Run(MS.Internal.CulturePreservingExecutionContext, System.Threading.ContextCallback, System.Object) at System.Windows.Threading.DispatcherOperation.Invoke() at System.Windows.Threading.Dispatcher.ProcessQueue() at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef) at MS.Win32.HwndWrapper.WndProc(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef) at MS.Win32.HwndSubclass.DispatcherCallbackOperation(System.Object) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate) at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(System.Windows.Threading.DispatcherPriority, System.TimeSpan, System.Delegate, System.Object, Int32) at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr, Int32, IntPtr, IntPtr) ```
True
Renamed an item from an import crashes VS - Renamed an item from an import ``` Application: devenv.exe Framework Version: v4.0.30319 Description: The process was terminated due to an unhandled exception. Exception Info: System.NullReferenceException at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<IsCompilationCaseSensitiveAsync>d__15.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task) at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<RenameAsync>d__13.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task) at System.Runtime.CompilerServices.TaskAwaiter.ValidateEnd(System.Threading.Tasks.Task) at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<OnWorkspaceChangedAsync>d__12.MoveNext() at System.Runtime.CompilerServices.AsyncMethodBuilderCore+<>c.<ThrowAsync>b__6_0(System.Object) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate) at System.Windows.Threading.DispatcherOperation.InvokeImpl() at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(System.Object) at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(System.Object) at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean) at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean) at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at MS.Internal.CulturePreservingExecutionContext.Run(MS.Internal.CulturePreservingExecutionContext, System.Threading.ContextCallback, System.Object) at System.Windows.Threading.DispatcherOperation.Invoke() at System.Windows.Threading.Dispatcher.ProcessQueue() at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef) at MS.Win32.HwndWrapper.WndProc(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef) at MS.Win32.HwndSubclass.DispatcherCallbackOperation(System.Object) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate) at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(System.Windows.Threading.DispatcherPriority, System.TimeSpan, System.Delegate, System.Object, Int32) at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr, Int32, IntPtr, IntPtr) ```
reli
renamed an item from an import crashes vs renamed an item from an import application devenv exe framework version description the process was terminated due to an unhandled exception exception info system nullreferenceexception at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices taskawaiter throwfornonsuccess system threading tasks task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification system threading tasks task at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices taskawaiter throwfornonsuccess system threading tasks task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification system threading tasks task at system runtime compilerservices taskawaiter validateend system threading tasks task at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices asyncmethodbuildercore c b system object at system windows threading exceptionwrapper internalrealcall system delegate system object at system windows threading exceptionwrapper trycatchwhen system object system delegate system object system delegate at system windows threading dispatcheroperation invokeimpl at system windows threading dispatcheroperation invokeinsecuritycontext system object at ms internal culturepreservingexecutioncontext callbackwrapper system object at system threading executioncontext runinternal system threading executioncontext system threading contextcallback system object boolean at system threading executioncontext run system threading executioncontext system threading contextcallback system object boolean at system threading executioncontext run system threading executioncontext system threading contextcallback system object at ms internal culturepreservingexecutioncontext run ms internal culturepreservingexecutioncontext system threading contextcallback system object at system windows threading dispatcheroperation invoke at system windows threading dispatcher processqueue at system windows threading dispatcher wndprochook intptr intptr intptr boolean byref at ms hwndwrapper wndproc intptr intptr intptr boolean byref at ms hwndsubclass dispatchercallbackoperation system object at system windows threading exceptionwrapper internalrealcall system delegate system object at system windows threading exceptionwrapper trycatchwhen system object system delegate system object system delegate at system windows threading dispatcher legacyinvokeimpl system windows threading dispatcherpriority system timespan system delegate system object at ms hwndsubclass subclasswndproc intptr intptr intptr
1
990
12,013,684,129
IssuesEvent
2020-04-10 09:27:25
microsoft/calculator
https://api.github.com/repos/microsoft/calculator
closed
AltGr + E causes unhandled exception 0x40080201
Area: Reliability Bug Pri: 2 triage approved
Calculator crashes when using the [`AltGr` key](https://en.wikipedia.org/wiki/AltGr_key) in combination with the `E` key. **Steps To Reproduce** 1. Open calculator in any mode (for example `Standard`). 2. Type nothing or anything you want. 3. Press at least once or hold the `AltGr` key on a physical or virtual keyboard. 4. Type nothing or anything you want, except the `E` key. 5. Press the `E` key. 6. Calculator terminates with unhandled exception code `0x40080201`. Pseudo-regex for the key presses after opening calc: `[^AltGr]* AltGr+ [^E]* E`. **Device and Application Information** - OS Build: 10.0.16299.0 - Architecture: x64 - Application Version: 10.1706.2406.0 - Region: de-DE - Dev Version Installed: False **Additional context** - Stacktrace: ``` KERNELBASE.dll!RaiseException+0x68 combase.dll!ObjectStublessClient32+0x5b443 combase.dll!RoOriginateError+0x51 vccorlib140_app.DLL!Platform::Details::EventSourceRemove+0x1239 vccorlib140_app.DLL!__abi_WinRTraiseInvalidCastException+0x17 vccorlib140_app.DLL!Platform::String::<Dispose>+0x1e8 vccorlib140_app.DLL!Platform::Details::ResolveWeakReference+0x98 Calculator.exe+0x3bd4e Calculator.exe+0x42b61 Windows.UI.dll!Ordinal1689+0x6e21 Windows.UI.dll!Ordinal1689+0x489e windows.ui.core.textinput.dll!Ordinal1501+0x31d0 windows.ui.core.textinput.dll!Ordinal1501+0x2a2b TextInputFramework.dll!TextInputHostSiteCreate+0x5dc1 TextInputFramework.dll!TextInputHostSiteCreate+0x82b0 TextInputFramework.dll!TextInputHostSiteCreate+0x5681 msctf.dll+0x3131 TextInputFramework.dll+0x18dc TextInputFramework.dll!TextInputHostCreateEx+0x1499f TextInputFramework.dll!TextInputHostCreateEx+0x14847 CoreMessaging.dll!CoreUICreateEx+0x2a8e CoreMessaging.dll!CoreUICallReceive+0x499 CoreMessaging.dll!CoreUICreateEx+0x12eef CoreMessaging.dll!CoreUICreateEx+0x1321a CoreMessaging.dll!CoreUICreateEx+0x1077b CoreMessaging.dll!CoreUICreateEx+0xbcc CoreMessaging.dll!CoreUICallReceive+0xca99 CoreMessaging.dll!CoreUICallReceive+0xcc88 CoreMessaging.dll!CoreUICallReceive+0xd186 CoreMessaging.dll!CoreUICallReceive+0xd4d4 CoreMessaging.dll!CoreUICreateEx+0x10a6a CoreMessaging.dll!CoreUICreateEx+0xffdc CoreMessaging.dll!CoreUICreateEx+0x7fbd CoreMessaging.dll!CoreUICreateEx+0x6527 CoreMessaging.dll!CoreUICreateEx+0x2d39 CoreMessaging.dll!CoreUICreateEx+0x2b19 user32.dll!CallWindowProcW+0x4dd user32.dll!CallWindowProcW+0x1cc user32.dll!GetTopWindow+0x153 ntdll.dll!KiUserCallbackDispatcher+0x24 win32u.dll!NtUserPeekMessage+0x14 user32.dll!PeekMessageW+0x88 Windows.UI.dll!Ordinal1500+0xd9f6 Windows.UI.dll!Ordinal1500+0xbbee Windows.UI.dll!Ordinal1500+0xd89b Windows.UI.Xaml.dll!DllGetActivationFactory+0x1856bc Windows.UI.Xaml.dll!DllGetActivationFactory+0x185660 twinapi.appcore.dll!Ordinal7+0x306a twinapi.appcore.dll!DllGetActivationFactory+0x146c shcore.dll!Ordinal247+0x1b5 KERNEL32.DLL!BaseThreadInitThunk+0x14 ntdll.dll!RtlUserThreadStart+0x21 ``` **Requested Assignment** I'm just reporting this problem. I don't want to fix it.
True
AltGr + E causes unhandled exception 0x40080201 - Calculator crashes when using the [`AltGr` key](https://en.wikipedia.org/wiki/AltGr_key) in combination with the `E` key. **Steps To Reproduce** 1. Open calculator in any mode (for example `Standard`). 2. Type nothing or anything you want. 3. Press at least once or hold the `AltGr` key on a physical or virtual keyboard. 4. Type nothing or anything you want, except the `E` key. 5. Press the `E` key. 6. Calculator terminates with unhandled exception code `0x40080201`. Pseudo-regex for the key presses after opening calc: `[^AltGr]* AltGr+ [^E]* E`. **Device and Application Information** - OS Build: 10.0.16299.0 - Architecture: x64 - Application Version: 10.1706.2406.0 - Region: de-DE - Dev Version Installed: False **Additional context** - Stacktrace: ``` KERNELBASE.dll!RaiseException+0x68 combase.dll!ObjectStublessClient32+0x5b443 combase.dll!RoOriginateError+0x51 vccorlib140_app.DLL!Platform::Details::EventSourceRemove+0x1239 vccorlib140_app.DLL!__abi_WinRTraiseInvalidCastException+0x17 vccorlib140_app.DLL!Platform::String::<Dispose>+0x1e8 vccorlib140_app.DLL!Platform::Details::ResolveWeakReference+0x98 Calculator.exe+0x3bd4e Calculator.exe+0x42b61 Windows.UI.dll!Ordinal1689+0x6e21 Windows.UI.dll!Ordinal1689+0x489e windows.ui.core.textinput.dll!Ordinal1501+0x31d0 windows.ui.core.textinput.dll!Ordinal1501+0x2a2b TextInputFramework.dll!TextInputHostSiteCreate+0x5dc1 TextInputFramework.dll!TextInputHostSiteCreate+0x82b0 TextInputFramework.dll!TextInputHostSiteCreate+0x5681 msctf.dll+0x3131 TextInputFramework.dll+0x18dc TextInputFramework.dll!TextInputHostCreateEx+0x1499f TextInputFramework.dll!TextInputHostCreateEx+0x14847 CoreMessaging.dll!CoreUICreateEx+0x2a8e CoreMessaging.dll!CoreUICallReceive+0x499 CoreMessaging.dll!CoreUICreateEx+0x12eef CoreMessaging.dll!CoreUICreateEx+0x1321a CoreMessaging.dll!CoreUICreateEx+0x1077b CoreMessaging.dll!CoreUICreateEx+0xbcc CoreMessaging.dll!CoreUICallReceive+0xca99 CoreMessaging.dll!CoreUICallReceive+0xcc88 CoreMessaging.dll!CoreUICallReceive+0xd186 CoreMessaging.dll!CoreUICallReceive+0xd4d4 CoreMessaging.dll!CoreUICreateEx+0x10a6a CoreMessaging.dll!CoreUICreateEx+0xffdc CoreMessaging.dll!CoreUICreateEx+0x7fbd CoreMessaging.dll!CoreUICreateEx+0x6527 CoreMessaging.dll!CoreUICreateEx+0x2d39 CoreMessaging.dll!CoreUICreateEx+0x2b19 user32.dll!CallWindowProcW+0x4dd user32.dll!CallWindowProcW+0x1cc user32.dll!GetTopWindow+0x153 ntdll.dll!KiUserCallbackDispatcher+0x24 win32u.dll!NtUserPeekMessage+0x14 user32.dll!PeekMessageW+0x88 Windows.UI.dll!Ordinal1500+0xd9f6 Windows.UI.dll!Ordinal1500+0xbbee Windows.UI.dll!Ordinal1500+0xd89b Windows.UI.Xaml.dll!DllGetActivationFactory+0x1856bc Windows.UI.Xaml.dll!DllGetActivationFactory+0x185660 twinapi.appcore.dll!Ordinal7+0x306a twinapi.appcore.dll!DllGetActivationFactory+0x146c shcore.dll!Ordinal247+0x1b5 KERNEL32.DLL!BaseThreadInitThunk+0x14 ntdll.dll!RtlUserThreadStart+0x21 ``` **Requested Assignment** I'm just reporting this problem. I don't want to fix it.
reli
altgr e causes unhandled exception calculator crashes when using the in combination with the e key steps to reproduce open calculator in any mode for example standard type nothing or anything you want press at least once or hold the altgr key on a physical or virtual keyboard type nothing or anything you want except the e key press the e key calculator terminates with unhandled exception code pseudo regex for the key presses after opening calc altgr e device and application information os build architecture application version region de de dev version installed false additional context stacktrace kernelbase dll raiseexception combase dll combase dll rooriginateerror app dll platform details eventsourceremove app dll abi winrtraiseinvalidcastexception app dll platform string app dll platform details resolveweakreference calculator exe calculator exe windows ui dll windows ui dll windows ui core textinput dll windows ui core textinput dll textinputframework dll textinputhostsitecreate textinputframework dll textinputhostsitecreate textinputframework dll textinputhostsitecreate msctf dll textinputframework dll textinputframework dll textinputhostcreateex textinputframework dll textinputhostcreateex coremessaging dll coreuicreateex coremessaging dll coreuicallreceive coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex dll callwindowprocw dll callwindowprocw dll gettopwindow ntdll dll kiusercallbackdispatcher dll ntuserpeekmessage dll peekmessagew windows ui dll windows ui dll windows ui dll windows ui xaml dll dllgetactivationfactory windows ui xaml dll dllgetactivationfactory twinapi appcore dll twinapi appcore dll dllgetactivationfactory shcore dll dll basethreadinitthunk ntdll dll rtluserthreadstart requested assignment i m just reporting this problem i don t want to fix it
1
14,504
2,814,133,886
IssuesEvent
2015-05-18 18:19:17
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[TEST-FAILURE] ClientMemberAttributeTest.testChangeMemberAttributes
Team: Client Type: Defect
``` java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds , count left: 1 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:603) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:596) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:588) at com.hazelcast.client.ClientMemberAttributeTest.testChangeMemberAttributes(ClientMemberAttributeTest.java:76) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client-new/405/testReport/junit/com.hazelcast.client/ClientMemberAttributeTest/testChangeMemberAttributes/
1.0
[TEST-FAILURE] ClientMemberAttributeTest.testChangeMemberAttributes - ``` java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds , count left: 1 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:603) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:596) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:588) at com.hazelcast.client.ClientMemberAttributeTest.testChangeMemberAttributes(ClientMemberAttributeTest.java:76) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client-new/405/testReport/junit/com.hazelcast.client/ClientMemberAttributeTest/testChangeMemberAttributes/
non_reli
clientmemberattributetest testchangememberattributes java lang assertionerror countdownlatch failed to complete within seconds count left at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast client clientmemberattributetest testchangememberattributes clientmemberattributetest java
0
236
5,530,185,596
IssuesEvent
2017-03-21 01:25:35
Azure/azure-webjobs-sdk-script
https://api.github.com/repos/Azure/azure-webjobs-sdk-script
closed
Scale controller needs more complex heuristic to estimate memory needs
reliability scale controller
There is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site, and the last writer wins. This results in situations where if one of the site’s instances is ‘inactive’, or under low load, let’s say due to inability to acquire some kind of lease, or being assigned to a non-busy partition, then, that worker will observe low a working set value, while a busy instance will have a (much) higher working set value. A better metric would be to keep the last ‘n’ readings (e.g. 5 or 10), and always use the highest. It’s still fairly simple, but should perform a lot better. Discussed in mail thread "app memory usage and instance scaling"
True
Scale controller needs more complex heuristic to estimate memory needs - There is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site, and the last writer wins. This results in situations where if one of the site’s instances is ‘inactive’, or under low load, let’s say due to inability to acquire some kind of lease, or being assigned to a non-busy partition, then, that worker will observe low a working set value, while a busy instance will have a (much) higher working set value. A better metric would be to keep the last ‘n’ readings (e.g. 5 or 10), and always use the highest. It’s still fairly simple, but should perform a lot better. Discussed in mail thread "app memory usage and instance scaling"
reli
scale controller needs more complex heuristic to estimate memory needs there is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site and the last writer wins this results in situations where if one of the site’s instances is ‘inactive’ or under low load let’s say due to inability to acquire some kind of lease or being assigned to a non busy partition then that worker will observe low a working set value while a busy instance will have a much higher working set value a better metric would be to keep the last ‘n’ readings e g or and always use the highest it’s still fairly simple but should perform a lot better discussed in mail thread app memory usage and instance scaling
1
187,857
14,433,311,848
IssuesEvent
2020-12-07 04:29:54
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
reopened
Windows LambdaLoadTest hang
test failure
ERROR: type should be string, got "https://ci.eclipse.org/openj9/job/Test_openjdk8_j9_special.system_x86-32_windows_Personal/33\r\nLambdaLoadTest_OpenJ9_NonLinux_special_24\r\nvariation: Mode107-OSRG\r\nJVM_OPTIONS: -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation \r\n\r\nNo diagnostic files generated.\r\n\r\n```\r\n10:22:03.554 - Completed 3.0%. Number of tests started=6\r\n10:22:23.882 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:22:43.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:03.413 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:03.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:23.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:43.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:23.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:43.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:03.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:03.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:23.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:43.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:23.508 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:23.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:23.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:43.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:23.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:43.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:03.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:03.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:43.400 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:23.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:23.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:03.400 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:43.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:22:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n```"
1.0
Windows LambdaLoadTest hang - https://ci.eclipse.org/openj9/job/Test_openjdk8_j9_special.system_x86-32_windows_Personal/33 LambdaLoadTest_OpenJ9_NonLinux_special_24 variation: Mode107-OSRG JVM_OPTIONS: -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation No diagnostic files generated. ``` 10:22:03.554 - Completed 3.0%. Number of tests started=6 10:22:23.882 - Completed 3.0%. Number of tests started=6 (+0) 10:22:43.460 - Completed 3.0%. Number of tests started=6 (+0) 10:23:03.476 - Completed 3.0%. Number of tests started=6 (+0) 10:23:23.492 - Completed 3.0%. Number of tests started=6 (+0) 10:23:43.398 - Completed 3.0%. Number of tests started=6 (+0) 10:24:03.413 - Completed 3.0%. Number of tests started=6 (+0) 10:24:23.429 - Completed 3.0%. Number of tests started=6 (+0) 10:24:43.445 - Completed 3.0%. Number of tests started=6 (+0) 10:25:03.460 - Completed 3.0%. Number of tests started=6 (+0) 10:25:23.476 - Completed 3.0%. Number of tests started=6 (+0) 10:25:43.492 - Completed 3.0%. Number of tests started=6 (+0) 10:26:03.398 - Completed 3.0%. Number of tests started=6 (+0) 10:26:23.414 - Completed 3.0%. Number of tests started=6 (+0) 10:26:43.429 - Completed 3.0%. Number of tests started=6 (+0) 10:27:03.445 - Completed 3.0%. Number of tests started=6 (+0) 10:27:23.460 - Completed 3.0%. Number of tests started=6 (+0) 10:27:43.476 - Completed 3.0%. Number of tests started=6 (+0) 10:28:03.492 - Completed 3.0%. Number of tests started=6 (+0) 10:28:23.398 - Completed 3.0%. Number of tests started=6 (+0) 10:28:43.414 - Completed 3.0%. Number of tests started=6 (+0) 10:29:03.429 - Completed 3.0%. Number of tests started=6 (+0) 10:29:23.445 - Completed 3.0%. Number of tests started=6 (+0) 10:29:43.461 - Completed 3.0%. Number of tests started=6 (+0) 10:30:03.476 - Completed 3.0%. Number of tests started=6 (+0) 10:30:23.492 - Completed 3.0%. Number of tests started=6 (+0) 10:30:43.398 - Completed 3.0%. Number of tests started=6 (+0) 10:31:03.414 - Completed 3.0%. Number of tests started=6 (+0) 10:31:23.429 - Completed 3.0%. Number of tests started=6 (+0) 10:31:43.445 - Completed 3.0%. Number of tests started=6 (+0) 10:32:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:32:23.414 - Completed 3.0%. Number of tests started=6 (+0) 10:32:43.398 - Completed 3.0%. Number of tests started=6 (+0) 10:33:03.429 - Completed 3.0%. Number of tests started=6 (+0) 10:33:23.445 - Completed 3.0%. Number of tests started=6 (+0) 10:33:43.461 - Completed 3.0%. Number of tests started=6 (+0) 10:34:03.476 - Completed 3.0%. Number of tests started=6 (+0) 10:34:23.492 - Completed 3.0%. Number of tests started=6 (+0) 10:34:43.398 - Completed 3.0%. Number of tests started=6 (+0) 10:35:03.414 - Completed 3.0%. Number of tests started=6 (+0) 10:35:23.429 - Completed 3.0%. Number of tests started=6 (+0) 10:35:43.445 - Completed 3.0%. Number of tests started=6 (+0) 10:36:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:36:23.476 - Completed 3.0%. Number of tests started=6 (+0) 10:36:43.492 - Completed 3.0%. Number of tests started=6 (+0) 10:37:03.398 - Completed 3.0%. Number of tests started=6 (+0) 10:37:23.414 - Completed 3.0%. Number of tests started=6 (+0) 10:37:43.430 - Completed 3.0%. Number of tests started=6 (+0) 10:38:03.445 - Completed 3.0%. Number of tests started=6 (+0) 10:38:23.461 - Completed 3.0%. Number of tests started=6 (+0) 10:38:43.476 - Completed 3.0%. Number of tests started=6 (+0) 10:39:03.492 - Completed 3.0%. Number of tests started=6 (+0) 10:39:23.398 - Completed 3.0%. Number of tests started=6 (+0) 10:39:43.414 - Completed 3.0%. Number of tests started=6 (+0) 10:40:03.430 - Completed 3.0%. Number of tests started=6 (+0) 10:40:23.445 - Completed 3.0%. Number of tests started=6 (+0) 10:40:43.461 - Completed 3.0%. Number of tests started=6 (+0) 10:41:03.476 - Completed 3.0%. Number of tests started=6 (+0) 10:41:23.492 - Completed 3.0%. Number of tests started=6 (+0) 10:41:43.398 - Completed 3.0%. Number of tests started=6 (+0) 10:42:03.414 - Completed 3.0%. Number of tests started=6 (+0) 10:42:23.430 - Completed 3.0%. Number of tests started=6 (+0) 10:42:43.445 - Completed 3.0%. Number of tests started=6 (+0) 10:43:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:43:23.477 - Completed 3.0%. Number of tests started=6 (+0) 10:43:43.492 - Completed 3.0%. Number of tests started=6 (+0) 10:44:03.398 - Completed 3.0%. Number of tests started=6 (+0) 10:44:23.414 - Completed 3.0%. Number of tests started=6 (+0) 10:44:43.430 - Completed 3.0%. Number of tests started=6 (+0) 10:45:03.445 - Completed 3.0%. Number of tests started=6 (+0) 10:45:23.461 - Completed 3.0%. Number of tests started=6 (+0) 10:45:43.477 - Completed 3.0%. Number of tests started=6 (+0) 10:46:03.492 - Completed 3.0%. Number of tests started=6 (+0) 10:46:23.398 - Completed 3.0%. Number of tests started=6 (+0) 10:46:43.414 - Completed 3.0%. Number of tests started=6 (+0) 10:47:03.430 - Completed 3.0%. Number of tests started=6 (+0) 10:47:23.508 - Completed 3.0%. Number of tests started=6 (+0) 10:47:43.430 - Completed 3.0%. Number of tests started=6 (+0) 10:48:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:48:23.399 - Completed 3.0%. Number of tests started=6 (+0) 10:48:43.414 - Completed 3.0%. Number of tests started=6 (+0) 10:49:03.430 - Completed 3.0%. Number of tests started=6 (+0) 10:49:23.446 - Completed 3.0%. Number of tests started=6 (+0) 10:49:43.461 - Completed 3.0%. Number of tests started=6 (+0) 10:50:03.477 - Completed 3.0%. Number of tests started=6 (+0) 10:50:23.492 - Completed 3.0%. Number of tests started=6 (+0) 10:50:43.399 - Completed 3.0%. Number of tests started=6 (+0) 10:51:03.414 - Completed 3.0%. Number of tests started=6 (+0) 10:51:23.430 - Completed 3.0%. Number of tests started=6 (+0) 10:51:43.446 - Completed 3.0%. Number of tests started=6 (+0) 10:52:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:52:23.477 - Completed 3.0%. Number of tests started=6 (+0) 10:52:43.492 - Completed 3.0%. Number of tests started=6 (+0) 10:53:03.399 - Completed 3.0%. Number of tests started=6 (+0) 10:53:23.414 - Completed 3.0%. Number of tests started=6 (+0) 10:53:43.430 - Completed 3.0%. Number of tests started=6 (+0) 10:54:03.446 - Completed 3.0%. Number of tests started=6 (+0) 10:54:23.461 - Completed 3.0%. Number of tests started=6 (+0) 10:54:43.477 - Completed 3.0%. Number of tests started=6 (+0) 10:55:03.493 - Completed 3.0%. Number of tests started=6 (+0) 10:55:23.399 - Completed 3.0%. Number of tests started=6 (+0) 10:55:43.414 - Completed 3.0%. Number of tests started=6 (+0) 10:56:03.430 - Completed 3.0%. Number of tests started=6 (+0) 10:56:23.446 - Completed 3.0%. Number of tests started=6 (+0) 10:56:43.461 - Completed 3.0%. Number of tests started=6 (+0) 10:57:03.477 - Completed 3.0%. Number of tests started=6 (+0) 10:57:23.493 - Completed 3.0%. Number of tests started=6 (+0) 10:57:43.399 - Completed 3.0%. Number of tests started=6 (+0) 10:58:03.414 - Completed 3.0%. Number of tests started=6 (+0) 10:58:23.430 - Completed 3.0%. Number of tests started=6 (+0) 10:58:43.446 - Completed 3.0%. Number of tests started=6 (+0) 10:59:03.461 - Completed 3.0%. Number of tests started=6 (+0) 10:59:23.477 - Completed 3.0%. Number of tests started=6 (+0) 10:59:43.493 - Completed 3.0%. Number of tests started=6 (+0) 11:00:03.399 - Completed 3.0%. Number of tests started=6 (+0) 11:00:23.415 - Completed 3.0%. Number of tests started=6 (+0) 11:00:43.430 - Completed 3.0%. Number of tests started=6 (+0) 11:01:03.446 - Completed 3.0%. Number of tests started=6 (+0) 11:01:23.461 - Completed 3.0%. Number of tests started=6 (+0) 11:01:43.477 - Completed 3.0%. Number of tests started=6 (+0) 11:02:03.493 - Completed 3.0%. Number of tests started=6 (+0) 11:02:23.399 - Completed 3.0%. Number of tests started=6 (+0) 11:02:43.415 - Completed 3.0%. Number of tests started=6 (+0) 11:03:03.430 - Completed 3.0%. Number of tests started=6 (+0) 11:03:23.446 - Completed 3.0%. Number of tests started=6 (+0) 11:03:43.462 - Completed 3.0%. Number of tests started=6 (+0) 11:04:03.477 - Completed 3.0%. Number of tests started=6 (+0) 11:04:23.493 - Completed 3.0%. Number of tests started=6 (+0) 11:04:43.399 - Completed 3.0%. Number of tests started=6 (+0) 11:05:03.415 - Completed 3.0%. Number of tests started=6 (+0) 11:05:23.430 - Completed 3.0%. Number of tests started=6 (+0) 11:05:43.446 - Completed 3.0%. Number of tests started=6 (+0) 11:06:03.462 - Completed 3.0%. Number of tests started=6 (+0) 11:06:23.477 - Completed 3.0%. Number of tests started=6 (+0) 11:06:43.493 - Completed 3.0%. Number of tests started=6 (+0) 11:07:03.399 - Completed 3.0%. Number of tests started=6 (+0) 11:07:23.415 - Completed 3.0%. Number of tests started=6 (+0) 11:07:43.430 - Completed 3.0%. Number of tests started=6 (+0) 11:08:03.446 - Completed 3.0%. Number of tests started=6 (+0) 11:08:23.462 - Completed 3.0%. Number of tests started=6 (+0) 11:08:43.477 - Completed 3.0%. Number of tests started=6 (+0) 11:09:03.493 - Completed 3.0%. Number of tests started=6 (+0) 11:09:23.399 - Completed 3.0%. Number of tests started=6 (+0) 11:09:43.415 - Completed 3.0%. Number of tests started=6 (+0) 11:10:03.430 - Completed 3.0%. Number of tests started=6 (+0) 11:10:23.446 - Completed 3.0%. Number of tests started=6 (+0) 11:10:43.462 - Completed 3.0%. Number of tests started=6 (+0) 11:11:03.477 - Completed 3.0%. Number of tests started=6 (+0) 11:11:23.493 - Completed 3.0%. Number of tests started=6 (+0) 11:11:43.399 - Completed 3.0%. Number of tests started=6 (+0) 11:12:03.415 - Completed 3.0%. Number of tests started=6 (+0) 11:12:23.431 - Completed 3.0%. Number of tests started=6 (+0) 11:12:43.446 - Completed 3.0%. Number of tests started=6 (+0) 11:13:03.462 - Completed 3.0%. Number of tests started=6 (+0) 11:13:23.477 - Completed 3.0%. Number of tests started=6 (+0) 11:13:43.493 - Completed 3.0%. Number of tests started=6 (+0) 11:14:03.399 - Completed 3.0%. Number of tests started=6 (+0) 11:14:23.415 - Completed 3.0%. Number of tests started=6 (+0) 11:14:43.431 - Completed 3.0%. Number of tests started=6 (+0) 11:15:03.446 - Completed 3.0%. Number of tests started=6 (+0) 11:15:23.462 - Completed 3.0%. Number of tests started=6 (+0) 11:15:43.478 - Completed 3.0%. Number of tests started=6 (+0) 11:16:03.493 - Completed 3.0%. Number of tests started=6 (+0) 11:16:23.399 - Completed 3.0%. Number of tests started=6 (+0) 11:16:43.415 - Completed 3.0%. Number of tests started=6 (+0) 11:17:03.431 - Completed 3.0%. Number of tests started=6 (+0) 11:17:23.446 - Completed 3.0%. Number of tests started=6 (+0) 11:17:43.462 - Completed 3.0%. Number of tests started=6 (+0) 11:18:03.478 - Completed 3.0%. Number of tests started=6 (+0) 11:18:23.493 - Completed 3.0%. Number of tests started=6 (+0) 11:18:43.400 - Completed 3.0%. Number of tests started=6 (+0) 11:19:03.415 - Completed 3.0%. Number of tests started=6 (+0) 11:19:23.431 - Completed 3.0%. Number of tests started=6 (+0) 11:19:43.446 - Completed 3.0%. Number of tests started=6 (+0) 11:20:03.462 - Completed 3.0%. Number of tests started=6 (+0) 11:20:23.478 - Completed 3.0%. Number of tests started=6 (+0) 11:20:43.493 - Completed 3.0%. Number of tests started=6 (+0) 11:21:03.400 - Completed 3.0%. Number of tests started=6 (+0) 11:21:23.415 - Completed 3.0%. Number of tests started=6 (+0) 11:21:43.431 - Completed 3.0%. Number of tests started=6 (+0) 11:22:03.446 - Completed 3.0%. Number of tests started=6 (+0) ```
non_reli
windows lambdaloadtest hang lambdaloadtest nonlinux special variation osrg jvm options xgcpolicy optthruput xdebug xrunjdwp transport dt socket address server y onthrow no pkg foo launch echo xjit enableosr enableosronguardfailure count disableasynccompilation no diagnostic files generated completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started
0
2,997
30,833,054,715
IssuesEvent
2023-08-02 04:29:16
hackforla/ops
https://api.github.com/repos/hackforla/ops
opened
[SPIKE] Set up AWS CloudWatch billing alarms
size: 2pt role: Site Reliability Engineer feature: monitoring
### Overview Currently, we have no safe guards in place if an AWS service costs more than anticipated. AWS has measures in place with tools like Cloudwatch. Let's investigate how to set up a cloudwatch alarm and see if we can set up safeguards. ### Action Items - [ ] Investigate HFLA's incubator AWS setup - [ ] Determine a general implementation and potential costs - [ ] Create implementation issue - [ ] Create CloudWatch documentation issue ### Resources/Instructions https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/monitor_estimated_charges_with_cloudwatch.html
True
[SPIKE] Set up AWS CloudWatch billing alarms - ### Overview Currently, we have no safe guards in place if an AWS service costs more than anticipated. AWS has measures in place with tools like Cloudwatch. Let's investigate how to set up a cloudwatch alarm and see if we can set up safeguards. ### Action Items - [ ] Investigate HFLA's incubator AWS setup - [ ] Determine a general implementation and potential costs - [ ] Create implementation issue - [ ] Create CloudWatch documentation issue ### Resources/Instructions https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/monitor_estimated_charges_with_cloudwatch.html
reli
set up aws cloudwatch billing alarms overview currently we have no safe guards in place if an aws service costs more than anticipated aws has measures in place with tools like cloudwatch let s investigate how to set up a cloudwatch alarm and see if we can set up safeguards action items investigate hfla s incubator aws setup determine a general implementation and potential costs create implementation issue create cloudwatch documentation issue resources instructions
1
466
7,698,594,608
IssuesEvent
2018-05-19 00:41:51
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
WinHttpRequestState memory leak when cancelling requests
area-System.Net.Http bug os-windows tenet-reliability
(reported by ASP.NET team). cc: @DavidObando There is a memory leak of WinHttpRequestState (and related objects) when the outstanding http request is canceled prior to it completing with an HttpResponseMessage.
True
WinHttpRequestState memory leak when cancelling requests - (reported by ASP.NET team). cc: @DavidObando There is a memory leak of WinHttpRequestState (and related objects) when the outstanding http request is canceled prior to it completing with an HttpResponseMessage.
reli
winhttprequeststate memory leak when cancelling requests reported by asp net team cc davidobando there is a memory leak of winhttprequeststate and related objects when the outstanding http request is canceled prior to it completing with an httpresponsemessage
1
2,851
28,236,391,837
IssuesEvent
2023-04-06 01:09:11
hackforla/ops
https://api.github.com/repos/hackforla/ops
opened
Provision RDS database for People Depot dev environment
size: 2pt role: Data Engineer role: Site Reliability Engineer feature: deployment
### Overview As a People Depot developer, I'd like an developer environment database provisioned for People Depot developers to work against. ### Action Items - [ ] Provision PostgresQL RDS - [ ] Give dev credentials to People Depot ### Resources/Instructions https://repost.aws/knowledge-center/ecs-fargate-task-database-connection
True
Provision RDS database for People Depot dev environment - ### Overview As a People Depot developer, I'd like an developer environment database provisioned for People Depot developers to work against. ### Action Items - [ ] Provision PostgresQL RDS - [ ] Give dev credentials to People Depot ### Resources/Instructions https://repost.aws/knowledge-center/ecs-fargate-task-database-connection
reli
provision rds database for people depot dev environment overview as a people depot developer i d like an developer environment database provisioned for people depot developers to work against action items provision postgresql rds give dev credentials to people depot resources instructions
1
1,109
13,146,873,446
IssuesEvent
2020-08-08 12:33:03
osbuild/osbuild-composer
https://api.github.com/repos/osbuild/osbuild-composer
opened
ci: exceeding openstack quota
ci-reliability
Now that our image tests spawn openstack instances we risk exceeding our quota, leading to test failure. A simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need.
True
ci: exceeding openstack quota - Now that our image tests spawn openstack instances we risk exceeding our quota, leading to test failure. A simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need.
reli
ci exceeding openstack quota now that our image tests spawn openstack instances we risk exceeding our quota leading to test failure a simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need
1
110,010
16,963,463,390
IssuesEvent
2021-06-29 08:03:47
elastic/kibana
https://api.github.com/repos/elastic/kibana
reopened
[Security Solution] Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West
QA:Validated Team: SecuritySolution Team:Threat Hunting bug v7.13.0
**Describe the feature** Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West **Build Details:** ``` Version: 7.12.1 BC3 Commit: d03a18355d794e508183a34734141626ce74b2c4 Build: 39452 Artifact Page: https://staging.elastic.co/7.12.1-7947c5bf/summary-7.12.1.html ``` **Preconditions** 1. Elastic 7.12.1 environment should be deployed 2. Alerts/Events should be generated **Steps to Reproduce** 1. Navigate to the Hosts Tab. 2. Go to the External alerts and open the analyzer for any alert/event 3. Observe that proper UI shows with nodes and information on the UI 4. Click on the North, South, East or West arrows until the node crosses the extremes of the UI allowed for the UI 5. Observe that the UI starts showing the message 'No Process Events Found'. **Test data** N/A **Impacted Test case(s)** N/A **Actual Result** Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West **Expected Result** Analyzer should show the nodes and data even if the user navigates to extreme North, South, East or West **What's Working** N/A **What's Not Working** N/A **Screenshot** - Normal analyzer ![Resolver](https://user-images.githubusercontent.com/60252716/115677821-7fc0bd80-a36e-11eb-8456-db7aee33e78c.JPG) - When the user navigates to extreme North, South, East or West ![Resolver1](https://user-images.githubusercontent.com/60252716/115677825-818a8100-a36e-11eb-9436-405cb5ab6955.JPG)
True
[Security Solution] Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West - **Describe the feature** Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West **Build Details:** ``` Version: 7.12.1 BC3 Commit: d03a18355d794e508183a34734141626ce74b2c4 Build: 39452 Artifact Page: https://staging.elastic.co/7.12.1-7947c5bf/summary-7.12.1.html ``` **Preconditions** 1. Elastic 7.12.1 environment should be deployed 2. Alerts/Events should be generated **Steps to Reproduce** 1. Navigate to the Hosts Tab. 2. Go to the External alerts and open the analyzer for any alert/event 3. Observe that proper UI shows with nodes and information on the UI 4. Click on the North, South, East or West arrows until the node crosses the extremes of the UI allowed for the UI 5. Observe that the UI starts showing the message 'No Process Events Found'. **Test data** N/A **Impacted Test case(s)** N/A **Actual Result** Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West **Expected Result** Analyzer should show the nodes and data even if the user navigates to extreme North, South, East or West **What's Working** N/A **What's Not Working** N/A **Screenshot** - Normal analyzer ![Resolver](https://user-images.githubusercontent.com/60252716/115677821-7fc0bd80-a36e-11eb-8456-db7aee33e78c.JPG) - When the user navigates to extreme North, South, East or West ![Resolver1](https://user-images.githubusercontent.com/60252716/115677825-818a8100-a36e-11eb-9436-405cb5ab6955.JPG)
non_reli
analyzer ui shows the message no process events found if the user navigates to extreme north south east or west describe the feature analyzer ui shows the message no process events found if the user navigates to extreme north south east or west build details version commit build artifact page preconditions elastic environment should be deployed alerts events should be generated steps to reproduce navigate to the hosts tab go to the external alerts and open the analyzer for any alert event observe that proper ui shows with nodes and information on the ui click on the north south east or west arrows until the node crosses the extremes of the ui allowed for the ui observe that the ui starts showing the message no process events found test data n a impacted test case s n a actual result analyzer ui shows the message no process events found if the user navigates to extreme north south east or west expected result analyzer should show the nodes and data even if the user navigates to extreme north south east or west what s working n a what s not working n a screenshot normal analyzer when the user navigates to extreme north south east or west
0
1,559
17,064,819,271
IssuesEvent
2021-07-07 05:32:50
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
Infinite loading and disconnecting issues
needs more info product-feedback reliability
Hello, Me and my team have been using this product for more than half a year now, but more than ever, it seems to be glitching everywhere and generally causing a lot of problems. Currently, we experience the following issues: changes don't save often times, the localhost is randomly failing, plus infinite or generally extremely slow loading speeds when simply changing a file or searching (file searching is a complete mess, it merely worked before). Not sure what we can do about this, since it's been increasingly frustrating the past couple of weeks, and it's not related to our server or our connection, which is through Wireless, same as before. Any help or advice would be welcome. Thanks!
True
Infinite loading and disconnecting issues - Hello, Me and my team have been using this product for more than half a year now, but more than ever, it seems to be glitching everywhere and generally causing a lot of problems. Currently, we experience the following issues: changes don't save often times, the localhost is randomly failing, plus infinite or generally extremely slow loading speeds when simply changing a file or searching (file searching is a complete mess, it merely worked before). Not sure what we can do about this, since it's been increasingly frustrating the past couple of weeks, and it's not related to our server or our connection, which is through Wireless, same as before. Any help or advice would be welcome. Thanks!
reli
infinite loading and disconnecting issues hello me and my team have been using this product for more than half a year now but more than ever it seems to be glitching everywhere and generally causing a lot of problems currently we experience the following issues changes don t save often times the localhost is randomly failing plus infinite or generally extremely slow loading speeds when simply changing a file or searching file searching is a complete mess it merely worked before not sure what we can do about this since it s been increasingly frustrating the past couple of weeks and it s not related to our server or our connection which is through wireless same as before any help or advice would be welcome thanks
1
752
10,348,015,677
IssuesEvent
2019-09-04 18:46:02
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Cancelling Http requests can result in "The handle is invalid" or other WinHTTP state errors
area-System.Net.Http bug os-windows tenet-reliability
Http requests can be cancelled in various ways using a cancellation token or even a timeout. Sometimes the CI tests timeout contacting the Azure test server due to network congestion, so this results in the requests getting cancelled as well by the HttpClient.Timeout property. When Http requests are cancelled in WinHttpHandler, the WinHTTP handle is closed. This cancels any outstanding async operations in WinHTTP. However, the current CancellationTokenRegistration code is closing the handle without first seeing if there is an active operation in flight. ``` c# private async void StartRequest(object obj) { // ... using (state.CancellationToken.Register(s => ((WinHttpRequestState)s).RequestHandle.Dispose(), state)) { do { // ... } while (state.RetryRequest); } ``` Since this cancellation is happening on a different thread from the main StartRequest() logic, it results in in-deterministic behavior. PR #8693 (Fix cancellation of WinHttpHandler response stream reads) has a similar CancellationToken.Register logic but it correctly checks for an active operation. This same logic needs to apply to the StartRequest() operation. This is the root cause of the following issues: #8928 #8496 #8426 #8061 #8000 ## History of failures ``` System.Net.Http.HttpRequestException : An error occurred while sending the request. ---- System.Net.Http.WinHttpException : The handle is invalid ``` Day | Build | OS | Test -- | -- | -- | -- 5/9 | 20170509.01 | Win7 | Timeout_SetTo30AndGetResponseFromLoopbackQuickly_Success 5/10 | 20170510.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 5/11 | 20170511.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 5/19 | 20170519.04 | Win7 | PostAsync_CallMethod_StreamContent 5/26 | 20170526.01 | Win7 | PostLongerContentLengths_UsesChunkedSemantics 5/30 | 20170530.06 | Win7 | SetDelegate_ConnectionSucceeds 6/4 | 20170604.02 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds 6/7 | 20170607.01 | Win7 | PostSyncBlockingContentUsingChunkedEncoding_Success 6/13 | 20170613.02 | Win7 | SetDelegate_ConnectionSucceeds 6/14 | 20170614.01 | Win7 | PostAsync_CallMethod_StreamContent 6/15 | 20170615.03 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds 6/15 | 20170615.03 | Win8.1 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized 6/16 | 20170616.01 | Win7 | PostAsync_CallMethod_EmptyContent 9/26 | 20170926.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 10/10 | 20171010.03 | Win7 | GetStreamAsync_ReadToEnd_Success 10/11 | 20171011.02 | Win7 | PostAsync_CallMethod_StreamContent 10/20 | 20171020.01 | Win7 | SetDelegate_ConnectionSucceeds 10/22 | 20171022.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent 10/22 | 20171022.02 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent 10/23 | 20171023.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 10/25 | 20171025.04 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 10/25 | 20171025.08 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 10/25 | 20171025.08 | Win7 | PostAsync_CallMethod_StreamContent 10/27 | 20171027.03 | Win7 | PostUsingContentLengthSemantics_Success 10/30 | 20171030.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 10/30 | 20171030.02 | Win7 | GetAsync_SupportedSSLVersion_Succeeds 10/31 | 20171030.05 | Win7 | SetDelegate_ConnectionSucceeds 11/3 | 20171103.07 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/6 | 20171106.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/6 | 20171106.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 11/7 | 20171107.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/9 | 20171109.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 11/15 | 20171115.02 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent 11/15 | 20171115.03 | Win7 | GetAsync_TrailingHeaders_Ignored 11/18 | 20171118.04 | Win7 | GetStreamAsync_ReadZeroBytes_Success 11/20 | 20171120.01 | Win7 | GetAsync_TrailingHeaders_Ignored 11/20 | 20171120.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 11/22 | 20171122.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/23 | 20171123.02 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized 11/24 | 20171124.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/27 | 20171127.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/28 | 20171128.01 | Win7 | SetDelegate_ConnectionSucceeds 11/28 | 20171128.07 | Win10 | UseCallback_NotSecureConnection_CallbackNotCalled 12/5 | 20171205.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 12/5 | 20171205.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 12/8 | 20171208.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent 12/12 | 20171212.02 | Win7 | GetStreamAsync_ReadToEnd_Success 12/23 | 20171223.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171223.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 12/24 | 20171224.02 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171224.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 12/31 | 20171231.01 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20171231.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized(authHeaders:%20%5C%22WWW-Authenticate:%20CustomAuth%5C%5Cr%5C%5Cn%5C%22)) 1/5 | 20180105.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180105.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True)) 1/10 | 20180110.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180110.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20201%20Created%5C%22,%20expectedStatusCode:%20201,%20expectedReason:%20%5C%22Created%5C%22)) 1/11 | 20180111.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180111.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20500%20Internal%20Server%20Error%5C%22,%20expectedStatusCode:%20500,%20expectedReason:%20%5C%22Inter)) 1/14 | 20180114.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180114.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostAsync_EmptyContent_ContentTypeHeaderNotSent(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/16 | 20180116.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180116.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20True)) 1/18 | 20180118.04 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180118.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False)) 1/19 | 20180119.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20180119.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test%7E2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20False)) 1/19 | 20180119.04 | Win7 | PostUsingContentLengthSemantics_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180119.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingContentLengthSemantics_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/21 | 20180121.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180121.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False)) 1/22 | 20180122.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180122.02/workItem/System.Net.Http.Functional.Tests/wilogs) 1/23 | 20180123.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRewindableContentUsingAuth_NoPreAuthenticate_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx%3Fauth=basic&user=user1&password=password1)) 1/23 | 20180123.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22)) 1/23 | 20180123.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 1/25 | 20180125.01 | Win7 | PostAsync_ReuseRequestContent_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180125.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FPostAsync_ReuseRequestContent_Success(remoteServer:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/27 | 20180127.01 | Win7 | GetAsync_TrailingHeaders_Ignored - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_TrailingHeaders_Ignored(includeTrailerHeader:%20True)) 1/27 | 20180127.04 | Win10 | PostRepeatedFlushContentUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRepeatedFlushContentUsingChunkedEncoding_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/27 | 20180127.04 | Win7 | GetAsync_RedirectResponseHasCookie_CookieSentToFinalUri - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/wilogs) 1/29 | 20180129.05 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180129.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 1/30 | 20180130.05 | Win7 | PostEmptyContentUsingConflictingSemantics_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180130.05/workItem/System.Net.Http.Functional.Tests/wilogs) 1/31 | 20180131.05 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180131.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True)) 2/1 | 20180201.01 | Win7 | PostUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingChunkedEncoding_Success(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 2/1 | 20180201.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 2/1 | 20180201.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22)) 2/5 | 20180205.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180205.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False)) 2/6 | 20180206.04 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False)) 2/6 | 20180206.05 | Win7 | UseCallback_NotSecureConnection_CallbackNotCalled - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_ServerCertificates_Test~2FUseCallback_NotSecureConnection_CallbackNotCalled) 2/8 | 20180208.01 | Win7 | 2x GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20O%20K%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22O%20K%5C%22)) and [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20600%20still%20valid%5C%22,%20expectedStatusCode:%20600,%20expectedReason:%20%5C%22still%20valid%5C%22)) 2/9 | 20180209.05 | Win7 | ThresholdExceeded_ThrowsException - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180209.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxResponseHeadersLength_Test~2FThresholdExceeded_ThrowsException(responseHeaders:%20%5C%22HTTP~2F1.1%20200%20OK%5C%5Cr%5C%5CnContent-Length:%200%5C%5Cr%5C%5CnCustom-000%5C%22...,%20maxResponseHeadersLe)) 2/10 | 20180210.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 2/10 | 20180210.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/wilogs) 2/13 | 20180213.01 | Win7 | ThresholdExceeded_ThrowsException 2/20 | 20180220.08 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success 2/21 | 20180221.01 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success 2/25 | 20180225.01 | Win7 | HttpClientHandler_IncorrectCredentials_Fails - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_IncorrectCredentials_Fails(authenticateHeader:%20%5C%22WWW-Authenticate:%20Digest%20realm=%5C%5C%5C%22hello%5C%5C%5C%22,%20nonce=%5C%5C%5C%22%5C%22...)) 2/25 | 20180225.02 | Win10 | HttpClientHandler_Authentication_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_Authentication_Succeeds(authenticateHeader:%20%5C%22Basic%20something,%20Digest%20something%5C%22,%20result:%20False)) 2/26 | 20180226.04 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180226.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test~2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20True)) 2/28 | 20180228.07 | Win7 | ThresholdExceeded_ThrowsException 3/1 | 20180301.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/2 | 20180302.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/4 | 20180304.01 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/7 | 20180307.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed 3/7 | 20180307.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/7 | 20180307.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/8 | 20180308.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/8 | 20180308.09 | Win7 | GetAsync_SetCookieContainer_CookieSent 3/10 | 20180310.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/14 | 20180314.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/17 | 20180317.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed 3/22 | 20180322.06 | Win7 | ThresholdExceeded_ThrowsException #### Similar failure with IOException wrapped in the middle ``` System.Net.Http.HttpRequestException : Error while copying content to a stream. ---- System.IO.IOException : The write operation failed, see inner exception. -------- System.Net.Http.WinHttpException : The handle is invalid ``` Day | Build | OS | Test -- | -- | -- | -- 11/17 | 20171117.04 | Win7 | PostUsingContentLengthSemantics_Success 12/2 | 20171202.01 | Win7 | PostAsync_ReuseRequestContent_Success
True
Cancelling Http requests can result in "The handle is invalid" or other WinHTTP state errors - Http requests can be cancelled in various ways using a cancellation token or even a timeout. Sometimes the CI tests timeout contacting the Azure test server due to network congestion, so this results in the requests getting cancelled as well by the HttpClient.Timeout property. When Http requests are cancelled in WinHttpHandler, the WinHTTP handle is closed. This cancels any outstanding async operations in WinHTTP. However, the current CancellationTokenRegistration code is closing the handle without first seeing if there is an active operation in flight. ``` c# private async void StartRequest(object obj) { // ... using (state.CancellationToken.Register(s => ((WinHttpRequestState)s).RequestHandle.Dispose(), state)) { do { // ... } while (state.RetryRequest); } ``` Since this cancellation is happening on a different thread from the main StartRequest() logic, it results in in-deterministic behavior. PR #8693 (Fix cancellation of WinHttpHandler response stream reads) has a similar CancellationToken.Register logic but it correctly checks for an active operation. This same logic needs to apply to the StartRequest() operation. This is the root cause of the following issues: #8928 #8496 #8426 #8061 #8000 ## History of failures ``` System.Net.Http.HttpRequestException : An error occurred while sending the request. ---- System.Net.Http.WinHttpException : The handle is invalid ``` Day | Build | OS | Test -- | -- | -- | -- 5/9 | 20170509.01 | Win7 | Timeout_SetTo30AndGetResponseFromLoopbackQuickly_Success 5/10 | 20170510.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 5/11 | 20170511.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 5/19 | 20170519.04 | Win7 | PostAsync_CallMethod_StreamContent 5/26 | 20170526.01 | Win7 | PostLongerContentLengths_UsesChunkedSemantics 5/30 | 20170530.06 | Win7 | SetDelegate_ConnectionSucceeds 6/4 | 20170604.02 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds 6/7 | 20170607.01 | Win7 | PostSyncBlockingContentUsingChunkedEncoding_Success 6/13 | 20170613.02 | Win7 | SetDelegate_ConnectionSucceeds 6/14 | 20170614.01 | Win7 | PostAsync_CallMethod_StreamContent 6/15 | 20170615.03 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds 6/15 | 20170615.03 | Win8.1 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized 6/16 | 20170616.01 | Win7 | PostAsync_CallMethod_EmptyContent 9/26 | 20170926.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 10/10 | 20171010.03 | Win7 | GetStreamAsync_ReadToEnd_Success 10/11 | 20171011.02 | Win7 | PostAsync_CallMethod_StreamContent 10/20 | 20171020.01 | Win7 | SetDelegate_ConnectionSucceeds 10/22 | 20171022.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent 10/22 | 20171022.02 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent 10/23 | 20171023.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 10/25 | 20171025.04 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 10/25 | 20171025.08 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 10/25 | 20171025.08 | Win7 | PostAsync_CallMethod_StreamContent 10/27 | 20171027.03 | Win7 | PostUsingContentLengthSemantics_Success 10/30 | 20171030.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 10/30 | 20171030.02 | Win7 | GetAsync_SupportedSSLVersion_Succeeds 10/31 | 20171030.05 | Win7 | SetDelegate_ConnectionSucceeds 11/3 | 20171103.07 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/6 | 20171106.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/6 | 20171106.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success 11/7 | 20171107.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/9 | 20171109.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 11/15 | 20171115.02 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent 11/15 | 20171115.03 | Win7 | GetAsync_TrailingHeaders_Ignored 11/18 | 20171118.04 | Win7 | GetStreamAsync_ReadZeroBytes_Success 11/20 | 20171120.01 | Win7 | GetAsync_TrailingHeaders_Ignored 11/20 | 20171120.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds 11/22 | 20171122.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/23 | 20171123.02 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized 11/24 | 20171124.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/27 | 20171127.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 11/28 | 20171128.01 | Win7 | SetDelegate_ConnectionSucceeds 11/28 | 20171128.07 | Win10 | UseCallback_NotSecureConnection_CallbackNotCalled 12/5 | 20171205.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 12/5 | 20171205.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 12/8 | 20171208.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent 12/12 | 20171212.02 | Win7 | GetStreamAsync_ReadToEnd_Success 12/23 | 20171223.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171223.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 12/24 | 20171224.02 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171224.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 12/31 | 20171231.01 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20171231.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized(authHeaders:%20%5C%22WWW-Authenticate:%20CustomAuth%5C%5Cr%5C%5Cn%5C%22)) 1/5 | 20180105.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180105.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True)) 1/10 | 20180110.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180110.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20201%20Created%5C%22,%20expectedStatusCode:%20201,%20expectedReason:%20%5C%22Created%5C%22)) 1/11 | 20180111.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180111.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20500%20Internal%20Server%20Error%5C%22,%20expectedStatusCode:%20500,%20expectedReason:%20%5C%22Inter)) 1/14 | 20180114.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180114.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostAsync_EmptyContent_ContentTypeHeaderNotSent(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/16 | 20180116.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180116.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20True)) 1/18 | 20180118.04 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180118.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False)) 1/19 | 20180119.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20180119.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test%7E2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20False)) 1/19 | 20180119.04 | Win7 | PostUsingContentLengthSemantics_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180119.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingContentLengthSemantics_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/21 | 20180121.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180121.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False)) 1/22 | 20180122.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180122.02/workItem/System.Net.Http.Functional.Tests/wilogs) 1/23 | 20180123.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRewindableContentUsingAuth_NoPreAuthenticate_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx%3Fauth=basic&user=user1&password=password1)) 1/23 | 20180123.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22)) 1/23 | 20180123.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 1/25 | 20180125.01 | Win7 | PostAsync_ReuseRequestContent_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180125.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FPostAsync_ReuseRequestContent_Success(remoteServer:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/27 | 20180127.01 | Win7 | GetAsync_TrailingHeaders_Ignored - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_TrailingHeaders_Ignored(includeTrailerHeader:%20True)) 1/27 | 20180127.04 | Win10 | PostRepeatedFlushContentUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRepeatedFlushContentUsingChunkedEncoding_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 1/27 | 20180127.04 | Win7 | GetAsync_RedirectResponseHasCookie_CookieSentToFinalUri - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/wilogs) 1/29 | 20180129.05 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180129.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True)) 1/30 | 20180130.05 | Win7 | PostEmptyContentUsingConflictingSemantics_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180130.05/workItem/System.Net.Http.Functional.Tests/wilogs) 1/31 | 20180131.05 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180131.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True)) 2/1 | 20180201.01 | Win7 | PostUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingChunkedEncoding_Success(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx)) 2/1 | 20180201.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 2/1 | 20180201.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22)) 2/5 | 20180205.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180205.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False)) 2/6 | 20180206.04 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False)) 2/6 | 20180206.05 | Win7 | UseCallback_NotSecureConnection_CallbackNotCalled - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_ServerCertificates_Test~2FUseCallback_NotSecureConnection_CallbackNotCalled) 2/8 | 20180208.01 | Win7 | 2x GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20O%20K%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22O%20K%5C%22)) and [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20600%20still%20valid%5C%22,%20expectedStatusCode:%20600,%20expectedReason:%20%5C%22still%20valid%5C%22)) 2/9 | 20180209.05 | Win7 | ThresholdExceeded_ThrowsException - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180209.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxResponseHeadersLength_Test~2FThresholdExceeded_ThrowsException(responseHeaders:%20%5C%22HTTP~2F1.1%20200%20OK%5C%5Cr%5C%5CnContent-Length:%200%5C%5Cr%5C%5CnCustom-000%5C%22...,%20maxResponseHeadersLe)) 2/10 | 20180210.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False)) 2/10 | 20180210.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/wilogs) 2/13 | 20180213.01 | Win7 | ThresholdExceeded_ThrowsException 2/20 | 20180220.08 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success 2/21 | 20180221.01 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success 2/25 | 20180225.01 | Win7 | HttpClientHandler_IncorrectCredentials_Fails - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_IncorrectCredentials_Fails(authenticateHeader:%20%5C%22WWW-Authenticate:%20Digest%20realm=%5C%5C%5C%22hello%5C%5C%5C%22,%20nonce=%5C%5C%5C%22%5C%22...)) 2/25 | 20180225.02 | Win10 | HttpClientHandler_Authentication_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_Authentication_Succeeds(authenticateHeader:%20%5C%22Basic%20something,%20Digest%20something%5C%22,%20result:%20False)) 2/26 | 20180226.04 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180226.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test~2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20True)) 2/28 | 20180228.07 | Win7 | ThresholdExceeded_ThrowsException 3/1 | 20180301.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/2 | 20180302.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/4 | 20180304.01 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/7 | 20180307.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed 3/7 | 20180307.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/7 | 20180307.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success 3/8 | 20180308.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/8 | 20180308.09 | Win7 | GetAsync_SetCookieContainer_CookieSent 3/10 | 20180310.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/14 | 20180314.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success 3/17 | 20180317.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed 3/22 | 20180322.06 | Win7 | ThresholdExceeded_ThrowsException #### Similar failure with IOException wrapped in the middle ``` System.Net.Http.HttpRequestException : Error while copying content to a stream. ---- System.IO.IOException : The write operation failed, see inner exception. -------- System.Net.Http.WinHttpException : The handle is invalid ``` Day | Build | OS | Test -- | -- | -- | -- 11/17 | 20171117.04 | Win7 | PostUsingContentLengthSemantics_Success 12/2 | 20171202.01 | Win7 | PostAsync_ReuseRequestContent_Success
reli
cancelling http requests can result in the handle is invalid or other winhttp state errors http requests can be cancelled in various ways using a cancellation token or even a timeout sometimes the ci tests timeout contacting the azure test server due to network congestion so this results in the requests getting cancelled as well by the httpclient timeout property when http requests are cancelled in winhttphandler the winhttp handle is closed this cancels any outstanding async operations in winhttp however the current cancellationtokenregistration code is closing the handle without first seeing if there is an active operation in flight c private async void startrequest object obj using state cancellationtoken register s winhttprequeststate s requesthandle dispose state do while state retryrequest since this cancellation is happening on a different thread from the main startrequest logic it results in in deterministic behavior pr fix cancellation of winhttphandler response stream reads has a similar cancellationtoken register logic but it correctly checks for an active operation this same logic needs to apply to the startrequest operation this is the root cause of the following issues history of failures system net http httprequestexception an error occurred while sending the request system net http winhttpexception the handle is invalid day build os test timeout success getasync allowedsslversion succeeds getasync allowedsslversion succeeds postasync callmethod streamcontent postlongercontentlengths useschunkedsemantics setdelegate connectionsucceeds nocallback revokedcertificate norevocationchecking succeeds postsyncblockingcontentusingchunkedencoding success setdelegate connectionsucceeds postasync callmethod streamcontent nocallback revokedcertificate norevocationchecking succeeds getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized postasync callmethod emptycontent readasstreamasync validserverresponse success getstreamasync readtoend success postasync callmethod streamcontent setdelegate connectionsucceeds maxresponsecontentbuffersize throwsiftoosmallforcontent maxresponsecontentbuffersize throwsiftoosmallforcontent readasstreamasync validserverresponse success getasync expectedstatuscodeandreason success postrewindablecontentusingauth nopreauthenticate success postasync callmethod streamcontent postusingcontentlengthsemantics success postrewindablecontentusingauth nopreauthenticate success getasync supportedsslversion succeeds setdelegate connectionsucceeds getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postrewindablecontentusingauth nopreauthenticate success getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success postasync emptycontent contenttypeheadernotsent getasync trailingheaders ignored getstreamasync readzerobytes success getasync trailingheaders ignored getasync allowedsslversion succeeds getasync expectedstatuscodeandreason success getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success setdelegate connectionsucceeds usecallback notsecureconnection callbacknotcalled getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postasync emptycontent contenttypeheadernotsent getstreamasync readtoend success getasync maxlimited concurrentcallsstillsucceed   getasync maxlimited concurrentcallsstillsucceed   getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized setdelegate connectionsucceeds getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postasync emptycontent contenttypeheadernotsent setdelegate connectionsucceeds maxresponsecontentbuffersize throwsiftoosmallforcontent getasync allowedsslversion succeeds postusingcontentlengthsemantics success maxresponsecontentbuffersize throwsiftoosmallforcontent readasstreamasync validserverresponse success only in postrewindablecontentusingauth nopreauthenticate success getasync expectedstatuscodeandreason success getasync maxlimited concurrentcallsstillsucceed postasync reuserequestcontent success getasync trailingheaders ignored postrepeatedflushcontentusingchunkedencoding success getasync redirectresponsehascookie cookiesenttofinaluri only in getasync maxlimited concurrentcallsstillsucceed postemptycontentusingconflictingsemantics success only in setdelegate connectionsucceeds postusingchunkedencoding success getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success setdelegate connectionsucceeds setdelegate connectionsucceeds usecallback notsecureconnection callbacknotcalled getasync expectedstatuscodeandreason success and thresholdexceeded throwsexception getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success only in thresholdexceeded throwsexception the handle is invalid getasync expectedstatuscodeandreason success the handle is invalid getasync expectedstatuscodeandreason success httpclienthandler incorrectcredentials fails httpclienthandler authentication succeeds getasync allowedsslversion succeeds thresholdexceeded throwsexception getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success readasstreamasync validserverresponse success getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success getasync expectedstatuscodeandreason success getasync setcookiecontainer cookiesent getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success getasync maxlimited concurrentcallsstillsucceed thresholdexceeded throwsexception similar failure with ioexception wrapped in the middle system net http httprequestexception error while copying content to a stream system io ioexception the write operation failed see inner exception system net http winhttpexception the handle is invalid day build os test postusingcontentlengthsemantics success postasync reuserequestcontent success
1
5,657
3,972,533,487
IssuesEvent
2016-05-04 15:36:21
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Mounting an AWS volume returns an error that looks incorrect: Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume
area/usability component/storage
This looks like an incorrect setup. Is this the dynamic provisioner? The error message is not useful to admins or users. ``` - apiVersion: v1 count: 1 firstTimestamp: 2016-05-04T14:33:04Z involvedObject: apiVersion: v1 kind: Pod name: jenkins-1-c1629 namespace: jwforres resourceVersion: "3983188" uid: fcddad4f-1204-11e6-98bd-0a1d348c34bb kind: Event lastTimestamp: 2016-05-04T14:33:04Z message: 'Unable to mount volumes for pod "jenkins-1-c1629_jwforres(fcddad4f-1204-11e6-98bd-0a1d348c34bb)": Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume (aws:///placeholder-for-provisioning)' metadata: creationTimestamp: 2016-05-04T14:33:04Z deletionTimestamp: 2016-05-04T16:33:04Z name: jenkins-1-c1629.144b63ef823f1dc1 namespace: jwforres resourceVersion: "3983383" selfLink: /api/v1/namespaces/jwforres/events/jenkins-1-c1629.144b63ef823f1dc1 uid: 1c7cf10f-1205-11e6-abb2-0ada84b8265d reason: FailedMount source: component: kubelet host: ip-172-31-14-20.ec2.internal type: Warning ```
True
Mounting an AWS volume returns an error that looks incorrect: Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume - This looks like an incorrect setup. Is this the dynamic provisioner? The error message is not useful to admins or users. ``` - apiVersion: v1 count: 1 firstTimestamp: 2016-05-04T14:33:04Z involvedObject: apiVersion: v1 kind: Pod name: jenkins-1-c1629 namespace: jwforres resourceVersion: "3983188" uid: fcddad4f-1204-11e6-98bd-0a1d348c34bb kind: Event lastTimestamp: 2016-05-04T14:33:04Z message: 'Unable to mount volumes for pod "jenkins-1-c1629_jwforres(fcddad4f-1204-11e6-98bd-0a1d348c34bb)": Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume (aws:///placeholder-for-provisioning)' metadata: creationTimestamp: 2016-05-04T14:33:04Z deletionTimestamp: 2016-05-04T16:33:04Z name: jenkins-1-c1629.144b63ef823f1dc1 namespace: jwforres resourceVersion: "3983383" selfLink: /api/v1/namespaces/jwforres/events/jenkins-1-c1629.144b63ef823f1dc1 uid: 1c7cf10f-1205-11e6-abb2-0ada84b8265d reason: FailedMount source: component: kubelet host: ip-172-31-14-20.ec2.internal type: Warning ```
non_reli
mounting an aws volume returns an error that looks incorrect could not attach ebs disk placeholder for provisioning invalid format for aws volume this looks like an incorrect setup is this the dynamic provisioner the error message is not useful to admins or users apiversion count firsttimestamp involvedobject apiversion kind pod name jenkins namespace jwforres resourceversion uid kind event lasttimestamp message unable to mount volumes for pod jenkins jwforres could not attach ebs disk placeholder for provisioning invalid format for aws volume aws placeholder for provisioning metadata creationtimestamp deletiontimestamp name jenkins namespace jwforres resourceversion selflink api namespaces jwforres events jenkins uid reason failedmount source component kubelet host ip internal type warning
0
2,932
30,316,765,236
IssuesEvent
2023-07-10 16:05:20
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Cancel on-going remote stream registration on stream removal
kind/bug area/performance area/reliability component/transport
**Describe the bug** There is currently a potential race condition which would result in a remote stream existing server side, even though the client stream has gone away. Since we register remote streams asynchronously, a remove request may be submitted client side, which will immediately remove it there. Then asynchronous removal requests are sent to the server. However, this can be interleaved with the asynchronous registration, resulting in a stream existing server side. The impact is additional latency during a push, or possible unnecessary job activation if it was the last stream for this type. However, the stream will eventually get removed appropriately. **Expected behavior** Registration/removal of remote streams is sequenced, such that a removal request would cancel registration attempts, and queue the removal after whatever in-flight requests were sent are finished. There is still a slight edge case around time outs, of course, but I think this is acceptable for now. The other option would be introducing even more coordination in the protocol, and I'd rather avoid this.
True
Cancel on-going remote stream registration on stream removal - **Describe the bug** There is currently a potential race condition which would result in a remote stream existing server side, even though the client stream has gone away. Since we register remote streams asynchronously, a remove request may be submitted client side, which will immediately remove it there. Then asynchronous removal requests are sent to the server. However, this can be interleaved with the asynchronous registration, resulting in a stream existing server side. The impact is additional latency during a push, or possible unnecessary job activation if it was the last stream for this type. However, the stream will eventually get removed appropriately. **Expected behavior** Registration/removal of remote streams is sequenced, such that a removal request would cancel registration attempts, and queue the removal after whatever in-flight requests were sent are finished. There is still a slight edge case around time outs, of course, but I think this is acceptable for now. The other option would be introducing even more coordination in the protocol, and I'd rather avoid this.
reli
cancel on going remote stream registration on stream removal describe the bug there is currently a potential race condition which would result in a remote stream existing server side even though the client stream has gone away since we register remote streams asynchronously a remove request may be submitted client side which will immediately remove it there then asynchronous removal requests are sent to the server however this can be interleaved with the asynchronous registration resulting in a stream existing server side the impact is additional latency during a push or possible unnecessary job activation if it was the last stream for this type however the stream will eventually get removed appropriately expected behavior registration removal of remote streams is sequenced such that a removal request would cancel registration attempts and queue the removal after whatever in flight requests were sent are finished there is still a slight edge case around time outs of course but i think this is acceptable for now the other option would be introducing even more coordination in the protocol and i d rather avoid this
1
468,217
13,463,030,587
IssuesEvent
2020-09-09 16:56:40
googleapis/releasetool
https://api.github.com/repos/googleapis/releasetool
closed
Magic proxy is currently failing, resulting in labels not being removed post publication
priority: p2 type: bug
#### Steps to reproduce 1. Merge a release PR, resulting in autorelease tagging a release. 2. It will enqueue a job which kicks off publication on kokoro. 3. When this job finishes, publication will succeed, but labels will not be released; the following error will be in logs: ```bash requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: https://magic-github-proxy.endpoints.devrel-prod.cloud.goog/repos/googleapis/java-document-ai/issues/133/comments? ```
1.0
Magic proxy is currently failing, resulting in labels not being removed post publication - #### Steps to reproduce 1. Merge a release PR, resulting in autorelease tagging a release. 2. It will enqueue a job which kicks off publication on kokoro. 3. When this job finishes, publication will succeed, but labels will not be released; the following error will be in logs: ```bash requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: https://magic-github-proxy.endpoints.devrel-prod.cloud.goog/repos/googleapis/java-document-ai/issues/133/comments? ```
non_reli
magic proxy is currently failing resulting in labels not being removed post publication steps to reproduce merge a release pr resulting in autorelease tagging a release it will enqueue a job which kicks off publication on kokoro when this job finishes publication will succeed but labels will not be released the following error will be in logs bash requests exceptions httperror client error bad request for url
0