Unnamed: 0 int64 9 832k | id float64 2.5B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 4 323 | labels stringlengths 4 2.67k | body stringlengths 23 107k | index stringclasses 4 values | text_combine stringlengths 96 107k | label stringclasses 2 values | text stringlengths 96 56.1k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,415 | 15,964,711,126 | IssuesEvent | 2021-04-16 06:42:13 | emmamei/cdkey | https://api.github.com/repos/emmamei/cdkey | opened | setDollType function doesn't check for valid type | bug reliabilityfix | When `setDollType` is called, it doesn't check for a valid Type. This needs to be fixed. | True | setDollType function doesn't check for valid type - When `setDollType` is called, it doesn't check for a valid Type. This needs to be fixed. | reli | setdolltype function doesn t check for valid type when setdolltype is called it doesn t check for a valid type this needs to be fixed | 1 |
2,267 | 24,562,137,570 | IssuesEvent | 2022-10-12 21:24:02 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | opened | [FEA] Should we synchronize and then spill with the ASYNC allocator | feature request ? - Needs Triage performance reliability | As mentioned here https://github.com/NVIDIA/spark-rapids/issues/6768, I am noticing the synchronizing on OOM can help us handle allocation failures that would otherwise be fatal. Additionally, with some quick prototyping locally, it seems that there may be a performance gain here.
Specifically, if we first `Cuda.deviceSynchronize` rather than spill right away, but fallback to the spill when we know we have already synchronized, we are able to save time with a quick query I tried in our performance cluster. I ran a query that spills constantly and it took 265 seconds vs 304 seconds without this change.
That said the query also ran OOM on a second trial. The reason I think is that we are really able to pack the GPU, I see that the async pool is able to get closer to its maximum size (40GB in this case). So we have less fudge memory for those tasks that run above their ~1/concurrentGpuTasks chunk of memory. | True | [FEA] Should we synchronize and then spill with the ASYNC allocator - As mentioned here https://github.com/NVIDIA/spark-rapids/issues/6768, I am noticing the synchronizing on OOM can help us handle allocation failures that would otherwise be fatal. Additionally, with some quick prototyping locally, it seems that there may be a performance gain here.
Specifically, if we first `Cuda.deviceSynchronize` rather than spill right away, but fallback to the spill when we know we have already synchronized, we are able to save time with a quick query I tried in our performance cluster. I ran a query that spills constantly and it took 265 seconds vs 304 seconds without this change.
That said the query also ran OOM on a second trial. The reason I think is that we are really able to pack the GPU, I see that the async pool is able to get closer to its maximum size (40GB in this case). So we have less fudge memory for those tasks that run above their ~1/concurrentGpuTasks chunk of memory. | reli | should we synchronize and then spill with the async allocator as mentioned here i am noticing the synchronizing on oom can help us handle allocation failures that would otherwise be fatal additionally with some quick prototyping locally it seems that there may be a performance gain here specifically if we first cuda devicesynchronize rather than spill right away but fallback to the spill when we know we have already synchronized we are able to save time with a quick query i tried in our performance cluster i ran a query that spills constantly and it took seconds vs seconds without this change that said the query also ran oom on a second trial the reason i think is that we are really able to pack the gpu i see that the async pool is able to get closer to its maximum size in this case so we have less fudge memory for those tasks that run above their concurrentgputasks chunk of memory | 1 |
1,645 | 18,050,901,285 | IssuesEvent | 2021-09-19 18:25:55 | ruuseljo/schoolprojectCO2 | https://api.github.com/repos/ruuseljo/schoolprojectCO2 | opened | the program should auto re-start if it has died | reliability | use a cron job to periodically check the processes for the main process. if it doesn't exist, re-start the process. if it does exist and is a zombie, kill and re-start. if it does exist and is not zombie, we are as happy as Larry. | True | the program should auto re-start if it has died - use a cron job to periodically check the processes for the main process. if it doesn't exist, re-start the process. if it does exist and is a zombie, kill and re-start. if it does exist and is not zombie, we are as happy as Larry. | reli | the program should auto re start if it has died use a cron job to periodically check the processes for the main process if it doesn t exist re start the process if it does exist and is a zombie kill and re start if it does exist and is not zombie we are as happy as larry | 1 |
454,154 | 13,095,710,533 | IssuesEvent | 2020-08-03 14:32:31 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Unit tests are broken | priority: medium quality | ## Describe the bug
Studio unit tests are broken in develop & support/3.1.x branchs
## To Reproduce
Steps to reproduce the behavior:
1. Clone studio develop branch or `./gradlew init`
2. Go into `src/studio` and run `mvn clean test`
3. See errors
## Expected behavior
All tests pass without error.
## Screenshots
N/A
## Logs
```
studio] mvn test
[INFO] Scanning for projects...
[INFO]
[INFO] -------------------< org.craftercms:crafter-studio >--------------------
[INFO] Building Crafter Studio 3.2.0-SNAPSHOT
[INFO] --------------------------------[ war ]---------------------------------
[INFO]
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ crafter-studio ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 218 resources
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create (Java Code) @ crafter-studio ---
[INFO] Checking for local modifications: skipped.
[INFO] Updating project files from SCM: skipped.
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio
[INFO] Storing buildNumber: a9ff303f84626f5ff74e51536c4be166a1aacd05 at timestamp: 1592499185453
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio
[INFO] Storing buildScmBranch: UNKNOWN
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (Java Code) @ crafter-studio ---
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create (UI-Version) @ crafter-studio ---
[INFO] Checking for local modifications: skipped.
[INFO] Updating project files from SCM: skipped.
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui
[WARNING] Cannot get the revision information from the scm repository, proceeding with revision of unknown :
Exception while executing SCM command.
[INFO] Storing buildNumber: unknown at timestamp: 1592499185630
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui
[WARNING] Cannot get the branch information from the scm repository, proceeding with UNKNOWN_BRANCH :
Exception while executing SCM command.
[INFO] Storing buildScmBranch: UNKNOWN_BRANCH
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (UI-Version) @ crafter-studio ---
[INFO]
[INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ crafter-studio ---
[INFO] Nothing to compile - all classes are up to date
[INFO]
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ crafter-studio ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 4 resources
[INFO]
[INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ crafter-studio ---
[INFO] Changes detected - recompiling the module!
[INFO] Compiling 5 source files to /home/jose/github/craftercms/src/studio/target/test-classes
[INFO] -------------------------------------------------------------
[ERROR] COMPILATION ERROR :
[INFO] -------------------------------------------------------------
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol
symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int)
location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[INFO] 5 errors
[INFO] -------------------------------------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.061 s
[INFO] Finished at: 2020-06-18T10:53:06-06:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project crafter-studio: Compilation failure: Compilation failure:
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol
[ERROR] symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int)
[ERROR] location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
```
## Specs
### Version
Studio Develop branch
Studio Support/3.1.x branch
### OS
Linux
### Browser
N/A
## Additional context
N/A
| 1.0 | [studio] Unit tests are broken - ## Describe the bug
Studio unit tests are broken in develop & support/3.1.x branchs
## To Reproduce
Steps to reproduce the behavior:
1. Clone studio develop branch or `./gradlew init`
2. Go into `src/studio` and run `mvn clean test`
3. See errors
## Expected behavior
All tests pass without error.
## Screenshots
N/A
## Logs
```
studio] mvn test
[INFO] Scanning for projects...
[INFO]
[INFO] -------------------< org.craftercms:crafter-studio >--------------------
[INFO] Building Crafter Studio 3.2.0-SNAPSHOT
[INFO] --------------------------------[ war ]---------------------------------
[INFO]
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ crafter-studio ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 218 resources
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create (Java Code) @ crafter-studio ---
[INFO] Checking for local modifications: skipped.
[INFO] Updating project files from SCM: skipped.
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio
[INFO] Storing buildNumber: a9ff303f84626f5ff74e51536c4be166a1aacd05 at timestamp: 1592499185453
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio
[INFO] Storing buildScmBranch: UNKNOWN
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (Java Code) @ crafter-studio ---
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create (UI-Version) @ crafter-studio ---
[INFO] Checking for local modifications: skipped.
[INFO] Updating project files from SCM: skipped.
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui
[WARNING] Cannot get the revision information from the scm repository, proceeding with revision of unknown :
Exception while executing SCM command.
[INFO] Storing buildNumber: unknown at timestamp: 1592499185630
[INFO] Executing: /bin/sh -c cd /home/jose/github/craftercms/src/studio/target/ui && git rev-parse --verify HEAD
[INFO] Working directory: /home/jose/github/craftercms/src/studio/target/ui
[WARNING] Cannot get the branch information from the scm repository, proceeding with UNKNOWN_BRANCH :
Exception while executing SCM command.
[INFO] Storing buildScmBranch: UNKNOWN_BRANCH
[INFO]
[INFO] --- buildnumber-maven-plugin:1.2:create-timestamp (UI-Version) @ crafter-studio ---
[INFO]
[INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ crafter-studio ---
[INFO] Nothing to compile - all classes are up to date
[INFO]
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ crafter-studio ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 4 resources
[INFO]
[INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ crafter-studio ---
[INFO] Changes detected - recompiling the module!
[INFO] Compiling 5 source files to /home/jose/github/craftercms/src/studio/target/test-classes
[INFO] -------------------------------------------------------------
[ERROR] COMPILATION ERROR :
[INFO] -------------------------------------------------------------
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol
symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol
symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int)
location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[INFO] 5 errors
[INFO] -------------------------------------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.061 s
[INFO] Finished at: 2020-06-18T10:53:06-06:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project crafter-studio: Compilation failure: Compilation failure:
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[257,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[275,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[299,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[302,22] cannot find symbol
[ERROR] symbol: method publish(java.lang.String,java.lang.String,java.util.List<org.craftercms.studio.api.v1.to.DeploymentItemTO>,java.lang.String,java.lang.String,java.lang.String)
[ERROR] location: class org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] /home/jose/github/craftercms/src/studio/src/test/java/org/craftercms/studio/impl/v2/repository/blob/BlobAwareContentRepositoryTest.java:[311,19] cannot find symbol
[ERROR] symbol: method getDeploymentHistory(java.lang.String,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,java.lang.Object,int)
[ERROR] location: variable local of type org.craftercms.studio.impl.v1.repository.git.GitContentRepository
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
```
## Specs
### Version
Studio Develop branch
Studio Support/3.1.x branch
### OS
Linux
### Browser
N/A
## Additional context
N/A
| non_reli | unit tests are broken describe the bug studio unit tests are broken in develop support x branchs to reproduce steps to reproduce the behavior clone studio develop branch or gradlew init go into src studio and run mvn clean test see errors expected behavior all tests pass without error screenshots n a logs studio mvn test scanning for projects building crafter studio snapshot maven resources plugin resources default resources crafter studio using utf encoding to copy filtered resources copying resources buildnumber maven plugin create java code crafter studio checking for local modifications skipped updating project files from scm skipped executing bin sh c cd home jose github craftercms src studio git rev parse verify head working directory home jose github craftercms src studio storing buildnumber at timestamp executing bin sh c cd home jose github craftercms src studio git rev parse verify head working directory home jose github craftercms src studio storing buildscmbranch unknown buildnumber maven plugin create timestamp java code crafter studio buildnumber maven plugin create ui version crafter studio checking for local modifications skipped updating project files from scm skipped executing bin sh c cd home jose github craftercms src studio target ui git rev parse verify head working directory home jose github craftercms src studio target ui cannot get the revision information from the scm repository proceeding with revision of unknown exception while executing scm command storing buildnumber unknown at timestamp executing bin sh c cd home jose github craftercms src studio target ui git rev parse verify head working directory home jose github craftercms src studio target ui cannot get the branch information from the scm repository proceeding with unknown branch exception while executing scm command storing buildscmbranch unknown branch buildnumber maven plugin create timestamp ui version crafter studio maven compiler plugin compile default compile crafter studio nothing to compile all classes are up to date maven resources plugin testresources default testresources crafter studio using utf encoding to copy filtered resources copying resources maven compiler plugin testcompile default testcompile crafter studio changes detected recompiling the module compiling source files to home jose github craftercms src studio target test classes compilation error home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method getdeploymenthistory java lang string java lang object java lang object java lang object java lang object java lang object int location variable local of type org craftercms studio impl repository git gitcontentrepository errors build failure total time s finished at failed to execute goal org apache maven plugins maven compiler plugin testcompile default testcompile on project crafter studio compilation failure compilation failure home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method publish java lang string java lang string java util list java lang string java lang string java lang string location class org craftercms studio impl repository git gitcontentrepository home jose github craftercms src studio src test java org craftercms studio impl repository blob blobawarecontentrepositorytest java cannot find symbol symbol method getdeploymenthistory java lang string java lang object java lang object java lang object java lang object java lang object int location variable local of type org craftercms studio impl repository git gitcontentrepository to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles specs version studio develop branch studio support x branch os linux browser n a additional context n a | 0 |
1,616 | 17,606,624,051 | IssuesEvent | 2021-08-17 17:57:58 | timberio/vector | https://api.github.com/repos/timberio/vector | opened | Running out of disk space should not stall disk buffering. | type: bug domain: buffers domain: reliability | Presently, if running a configuration with disk buffers enabled, the process of running out of space will result in a panic being thrown in `lib/vector-core/buffers/src/disk/leveldb_buffer/writer.rs:189` due to unwrapping a result which ends up being the I/O error and not the expected success of the write.
Since disk buffering will correctly stop buffering once we hit the configured `buffer.max_size`, we should consider running out of disk space to be approximately the same condition. This might be something we can only generally capture, given that I don't know if there's a way, at the type level, to assert that an I/O error is specifically related to running out of disk space. We'd be doing a best effort check on the error's description.
One might also argue that an I/O error in disk buffering on the write path should simply continue along, applying normal backpressure, but that's a bigger discussion. | True | Running out of disk space should not stall disk buffering. - Presently, if running a configuration with disk buffers enabled, the process of running out of space will result in a panic being thrown in `lib/vector-core/buffers/src/disk/leveldb_buffer/writer.rs:189` due to unwrapping a result which ends up being the I/O error and not the expected success of the write.
Since disk buffering will correctly stop buffering once we hit the configured `buffer.max_size`, we should consider running out of disk space to be approximately the same condition. This might be something we can only generally capture, given that I don't know if there's a way, at the type level, to assert that an I/O error is specifically related to running out of disk space. We'd be doing a best effort check on the error's description.
One might also argue that an I/O error in disk buffering on the write path should simply continue along, applying normal backpressure, but that's a bigger discussion. | reli | running out of disk space should not stall disk buffering presently if running a configuration with disk buffers enabled the process of running out of space will result in a panic being thrown in lib vector core buffers src disk leveldb buffer writer rs due to unwrapping a result which ends up being the i o error and not the expected success of the write since disk buffering will correctly stop buffering once we hit the configured buffer max size we should consider running out of disk space to be approximately the same condition this might be something we can only generally capture given that i don t know if there s a way at the type level to assert that an i o error is specifically related to running out of disk space we d be doing a best effort check on the error s description one might also argue that an i o error in disk buffering on the write path should simply continue along applying normal backpressure but that s a bigger discussion | 1 |
578,914 | 17,156,555,898 | IssuesEvent | 2021-07-14 07:42:01 | vrchatapi/specification | https://api.github.com/repos/vrchatapi/specification | closed | Issue with accountDeletionDate in CurrentUser.yaml | Priority: Medium Status: Accepted Type: Maintenance | ```
accountDeletionDate:
oneOf:
- $ref: ./Date.yaml
- nullable: true
```
I think something is wrong with this piece but I don't know enough about the OpenAPI specification to fix it.
Here are some links I've found regarding this:
https://github.com/Redocly/redoc/issues/1303
https://stackoverflow.com/questions/40920441/how-to-specify-a-property-can-be-null-or-a-reference-with-swagger
I tried doing some of the things in those links but I ended up just making it worse | 1.0 | Issue with accountDeletionDate in CurrentUser.yaml - ```
accountDeletionDate:
oneOf:
- $ref: ./Date.yaml
- nullable: true
```
I think something is wrong with this piece but I don't know enough about the OpenAPI specification to fix it.
Here are some links I've found regarding this:
https://github.com/Redocly/redoc/issues/1303
https://stackoverflow.com/questions/40920441/how-to-specify-a-property-can-be-null-or-a-reference-with-swagger
I tried doing some of the things in those links but I ended up just making it worse | non_reli | issue with accountdeletiondate in currentuser yaml accountdeletiondate oneof ref date yaml nullable true i think something is wrong with this piece but i don t know enough about the openapi specification to fix it here are some links i ve found regarding this i tried doing some of the things in those links but i ended up just making it worse | 0 |
3,033 | 31,778,268,082 | IssuesEvent | 2023-09-12 15:38:38 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | closed | [FEA] SupportSplitAndRetry for GpuRangeExec | feature request reliability | **Is your feature request related to a problem? Please describe.**
GpuRangeExec is mostly done on the CPU. Only once it has a start, end, and step for a batch does it move onto the GPU for processing.
The main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch.
For split and retry we would need to make sure that if we see a split that we can drop the number of `rowsThisBatch`, but not too small.
| True | [FEA] SupportSplitAndRetry for GpuRangeExec - **Is your feature request related to a problem? Please describe.**
GpuRangeExec is mostly done on the CPU. Only once it has a start, end, and step for a batch does it move onto the GPU for processing.
The main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch.
For split and retry we would need to make sure that if we see a split that we can drop the number of `rowsThisBatch`, but not too small.
| reli | supportsplitandretry for gpurangeexec is your feature request related to a problem please describe gpurangeexec is mostly done on the cpu only once it has a start end and step for a batch does it move onto the gpu for processing the main changes here for retry would be to refactor the code so that we update the global state after we have generated the output batch for split and retry we would need to make sure that if we see a split that we can drop the number of rowsthisbatch but not too small | 1 |
576 | 8,689,972,313 | IssuesEvent | 2018-12-03 20:12:08 | rook/rook | https://api.github.com/repos/rook/rook | closed | Proper cleanup after lost nodes/OSDs | operator reliability wontfix | * Bug Report
**Deviation from expected behavior:**
Configmaps and deployments related to terminated k8s nodes are left dangling, and OSDs are not taken out of the ceph cluster.
**Expected behavior:**
I would expect that configmaps and deployments for deleted/terminated/missing nodes to be removed by the operator.
**How to reproduce it (minimal and precise):**
Create a multi node K8s cluster, and add a OSDs on each node. Terminate/delete one of the k8s nodes and watch that no cleanup is not performed.
To cleanup you have to
1. Delete the job and configmap in the rook-ceph namespace related the dead/removed node
2. Delete the deployment in the rook-ceph namespace related to the osd that was present on the dead/removed node
3. Delete the two configmaps related to the dead/removed node in the rook-ceph-system namespace
4. Restart the rook operator.
Also dead OSDs should be removed from the ceph cluster by issuing `ceph osd purge osd.ID --yes-i-really-mean-it`, where ID is the OSD that was previously running on the terminated node.
Refer to this thread on slack for more details
https://rook-io.slack.com/messages/C46Q5UC05/convo/C46Q5UC05-1534330458.000100/
**Environment**:
* OS (e.g. from /etc/os-release): CoreOS
* Cloud provider or hardware configuration: AWS
* Rook version (use `rook version` inside of a Rook Pod): 0.8.1
* Kubernetes version (use `kubectl version`): 1.9.3
* Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift): On AWS, built with kube-aws
* Storage backend status (e.g. for Ceph use `ceph health` in the [Rook Ceph toolbox](https://rook.io/docs/Rook/master/toolbox.html)):
OK | True | Proper cleanup after lost nodes/OSDs - * Bug Report
**Deviation from expected behavior:**
Configmaps and deployments related to terminated k8s nodes are left dangling, and OSDs are not taken out of the ceph cluster.
**Expected behavior:**
I would expect that configmaps and deployments for deleted/terminated/missing nodes to be removed by the operator.
**How to reproduce it (minimal and precise):**
Create a multi node K8s cluster, and add a OSDs on each node. Terminate/delete one of the k8s nodes and watch that no cleanup is not performed.
To cleanup you have to
1. Delete the job and configmap in the rook-ceph namespace related the dead/removed node
2. Delete the deployment in the rook-ceph namespace related to the osd that was present on the dead/removed node
3. Delete the two configmaps related to the dead/removed node in the rook-ceph-system namespace
4. Restart the rook operator.
Also dead OSDs should be removed from the ceph cluster by issuing `ceph osd purge osd.ID --yes-i-really-mean-it`, where ID is the OSD that was previously running on the terminated node.
Refer to this thread on slack for more details
https://rook-io.slack.com/messages/C46Q5UC05/convo/C46Q5UC05-1534330458.000100/
**Environment**:
* OS (e.g. from /etc/os-release): CoreOS
* Cloud provider or hardware configuration: AWS
* Rook version (use `rook version` inside of a Rook Pod): 0.8.1
* Kubernetes version (use `kubectl version`): 1.9.3
* Kubernetes cluster type (e.g. Tectonic, GKE, OpenShift): On AWS, built with kube-aws
* Storage backend status (e.g. for Ceph use `ceph health` in the [Rook Ceph toolbox](https://rook.io/docs/Rook/master/toolbox.html)):
OK | reli | proper cleanup after lost nodes osds bug report deviation from expected behavior configmaps and deployments related to terminated nodes are left dangling and osds are not taken out of the ceph cluster expected behavior i would expect that configmaps and deployments for deleted terminated missing nodes to be removed by the operator how to reproduce it minimal and precise create a multi node cluster and add a osds on each node terminate delete one of the nodes and watch that no cleanup is not performed to cleanup you have to delete the job and configmap in the rook ceph namespace related the dead removed node delete the deployment in the rook ceph namespace related to the osd that was present on the dead removed node delete the two configmaps related to the dead removed node in the rook ceph system namespace restart the rook operator also dead osds should be removed from the ceph cluster by issuing ceph osd purge osd id yes i really mean it where id is the osd that was previously running on the terminated node refer to this thread on slack for more details environment os e g from etc os release coreos cloud provider or hardware configuration aws rook version use rook version inside of a rook pod kubernetes version use kubectl version kubernetes cluster type e g tectonic gke openshift on aws built with kube aws storage backend status e g for ceph use ceph health in the ok | 1 |
757 | 10,476,136,405 | IssuesEvent | 2019-09-23 17:54:25 | microsoft/BotFramework-DirectLineJS | https://api.github.com/repos/microsoft/BotFramework-DirectLineJS | opened | Happy path: with conversation ID, typing activity can be sent | Reliability Streaming Extensions | 1. Start a conversation by specifying a conversation ID
1. Send a typing activity
1. Make sure the bot can receive it | True | Happy path: with conversation ID, typing activity can be sent - 1. Start a conversation by specifying a conversation ID
1. Send a typing activity
1. Make sure the bot can receive it | reli | happy path with conversation id typing activity can be sent start a conversation by specifying a conversation id send a typing activity make sure the bot can receive it | 1 |
2,328 | 24,711,085,152 | IssuesEvent | 2022-10-20 00:51:59 | ppy/osu | https://api.github.com/repos/ppy/osu | closed | Game crashing after map loaded from the editor. | type:online type:reliability | ### Type
Crash to desktop
### Bug description
I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one.
The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing.
5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black.
**The game crashed** after like **3 seconds**.
This stopped happening after I wrote this and I still don't know what caused this reactive behavior.
### Screenshots or videos
I don't have any photos/videos that show this in action.
### Version
osu! 2022.1008.2-lazer
### Logs
I'm going to send them all, but I found that runtime is most interesting...
[database.log](https://github.com/ppy/osu/files/9785324/database.log)
[input.log](https://github.com/ppy/osu/files/9785326/input.log)
[network.log](https://github.com/ppy/osu/files/9785327/network.log)
[performance.log](https://github.com/ppy/osu/files/9785328/performance.log)
[runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log)
[updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
| True | Game crashing after map loaded from the editor. - ### Type
Crash to desktop
### Bug description
I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one.
The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing.
5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black.
**The game crashed** after like **3 seconds**.
This stopped happening after I wrote this and I still don't know what caused this reactive behavior.
### Screenshots or videos
I don't have any photos/videos that show this in action.
### Version
osu! 2022.1008.2-lazer
### Logs
I'm going to send them all, but I found that runtime is most interesting...
[database.log](https://github.com/ppy/osu/files/9785324/database.log)
[input.log](https://github.com/ppy/osu/files/9785326/input.log)
[network.log](https://github.com/ppy/osu/files/9785327/network.log)
[performance.log](https://github.com/ppy/osu/files/9785328/performance.log)
[runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log)
[updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
| reli | game crashing after map loaded from the editor type crash to desktop bug description i was using osu without an internet connection and wanted to play some maps i got to the song selector and loaded a map but it was taking a lot of time so i canceled the map and reloaded another one the same thing happened so i loaded another map in the editor to see if there was something wrong but there wasn t so i played the map from the editor this time everything was going well but after seconds of playing the game froze but the audio was still playing seconds later the screen went black but the audio was still playing seconds after that the audio stopped playing and the usual osu lazer has stopped working popped up but i was waiting then the game played audio again but the screen was still black the game crashed after like seconds this stopped happening after i wrote this and i still don t know what caused this reactive behavior screenshots or videos i don t have any photos videos that show this in action version osu lazer logs i m going to send them all but i found that runtime is most interesting | 1 |
237,972 | 7,768,754,195 | IssuesEvent | 2018-06-03 21:27:48 | TerraFusion/basicFusion | https://api.github.com/repos/TerraFusion/basicFusion | closed | Modify generation scripts for BF output directories | Medium Priority enhancement | Modify the generation scripts so that the BF output files are stored in some sort of directory hierarchy based on dates. | 1.0 | Modify generation scripts for BF output directories - Modify the generation scripts so that the BF output files are stored in some sort of directory hierarchy based on dates. | non_reli | modify generation scripts for bf output directories modify the generation scripts so that the bf output files are stored in some sort of directory hierarchy based on dates | 0 |
317 | 6,558,869,926 | IssuesEvent | 2017-09-06 23:48:32 | waggle-sensor/beehive-server | https://api.github.com/repos/waggle-sensor/beehive-server | opened | Prototype static version of beehive | reliability | One _possible_ improvement we can do is build a static version of beehive which is regenerated on a schedule. This would dramatically improve page serving performance across the board. This also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up.
I think this is still worth prototyping, even though we now have nginx performing caching and have moved off the development server. As an example, the build-index tool in the data-exporter generates a "friendly" summary of all the datasets to make sure things look reasonable.
| True | Prototype static version of beehive - One _possible_ improvement we can do is build a static version of beehive which is regenerated on a schedule. This would dramatically improve page serving performance across the board. This also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up.
I think this is still worth prototyping, even though we now have nginx performing caching and have moved off the development server. As an example, the build-index tool in the data-exporter generates a "friendly" summary of all the datasets to make sure things look reasonable.
| reli | prototype static version of beehive one possible improvement we can do is build a static version of beehive which is regenerated on a schedule this would dramatically improve page serving performance across the board this also has the side effect of completely eliminating direct database access for datasets to the outside work and so could eliminate any security mistakes which show up i think this is still worth prototyping even though we now have nginx performing caching and have moved off the development server as an example the build index tool in the data exporter generates a friendly summary of all the datasets to make sure things look reasonable | 1 |
13,067 | 8,788,279,540 | IssuesEvent | 2018-12-20 21:36:13 | istio/istio | https://api.github.com/repos/istio/istio | closed | Citadel in CrashLoopBackOff | area/security stale | **Describe the bug**
Use helm to create a yaml file including grafana
**Expected behavior**
running grafana container
**Steps to reproduce the bug**
helm template ../istio/install/kubernetes/helm/istio --name istio --namespace istio-system --set grafana.enabled=true --set grafana.security.enabled=true --set grafana.security.adminUser=admin --set grafana.security.adminPassword=mypass > istio.yaml
kubectl create ns istio-system
kubectl apply -n istio-system -f istio.yaml
```
$ kubectl get all --all-namespaces | grep -i istio
istio-system pod/grafana-67c5b957f9-l79mq 0/1 CreateContainerConfigError 0 8m
istio-system pod/istio-citadel-7d9c98786d-prpwn 0/1 CrashLoopBackOff 6 8m
istio-system pod/istio-cleanup-secrets-s5fhb 0/1 Completed 0 8m
istio-system pod/istio-egressgateway-7bdd8c695f-t6g6f 1/1 Running 0 8m
istio-system pod/istio-galley-6c799667d4-jscww 0/1 ContainerCreating 0 8m
...
istio-system pod/istio-sidecar-injector-c8c4c568b-55h7z 0/1 ContainerCreating 0 8m
...
```
**Version**
```
$ istioctl version
Version: 1.0.1
GitRevision: 42773aacced474d97159902d20579a25b1f98106
User: root@832d5020b1d4
Hub: gcr.io/istio-release
GolangVersion: go1.10.1
BuildStatus: Clean
$ kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.5", GitCommit:"32ac1c9073b132b8ba18aa830f46b77dcceb0723", GitTreeState:"clean", BuildDate:"2018-06-21T11:34:22Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"}
```
**Is Istio Auth enabled or not?**
No idea
**Environment**
Kops w/coreOS on AWS
**Cluster state**
[istio-dump.tar.gz](https://github.com/istio/istio/files/2358972/istio-dump.tar.gz)
| True | Citadel in CrashLoopBackOff - **Describe the bug**
Use helm to create a yaml file including grafana
**Expected behavior**
running grafana container
**Steps to reproduce the bug**
helm template ../istio/install/kubernetes/helm/istio --name istio --namespace istio-system --set grafana.enabled=true --set grafana.security.enabled=true --set grafana.security.adminUser=admin --set grafana.security.adminPassword=mypass > istio.yaml
kubectl create ns istio-system
kubectl apply -n istio-system -f istio.yaml
```
$ kubectl get all --all-namespaces | grep -i istio
istio-system pod/grafana-67c5b957f9-l79mq 0/1 CreateContainerConfigError 0 8m
istio-system pod/istio-citadel-7d9c98786d-prpwn 0/1 CrashLoopBackOff 6 8m
istio-system pod/istio-cleanup-secrets-s5fhb 0/1 Completed 0 8m
istio-system pod/istio-egressgateway-7bdd8c695f-t6g6f 1/1 Running 0 8m
istio-system pod/istio-galley-6c799667d4-jscww 0/1 ContainerCreating 0 8m
...
istio-system pod/istio-sidecar-injector-c8c4c568b-55h7z 0/1 ContainerCreating 0 8m
...
```
**Version**
```
$ istioctl version
Version: 1.0.1
GitRevision: 42773aacced474d97159902d20579a25b1f98106
User: root@832d5020b1d4
Hub: gcr.io/istio-release
GolangVersion: go1.10.1
BuildStatus: Clean
$ kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"bb9ffb1654d4a729bb4cec18ff088eacc153c239", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"10", GitVersion:"v1.10.5", GitCommit:"32ac1c9073b132b8ba18aa830f46b77dcceb0723", GitTreeState:"clean", BuildDate:"2018-06-21T11:34:22Z", GoVersion:"go1.9.3", Compiler:"gc", Platform:"linux/amd64"}
```
**Is Istio Auth enabled or not?**
No idea
**Environment**
Kops w/coreOS on AWS
**Cluster state**
[istio-dump.tar.gz](https://github.com/istio/istio/files/2358972/istio-dump.tar.gz)
| non_reli | citadel in crashloopbackoff describe the bug use helm to create a yaml file including grafana expected behavior running grafana container steps to reproduce the bug helm template istio install kubernetes helm istio name istio namespace istio system set grafana enabled true set grafana security enabled true set grafana security adminuser admin set grafana security adminpassword mypass istio yaml kubectl create ns istio system kubectl apply n istio system f istio yaml kubectl get all all namespaces grep i istio istio system pod grafana createcontainerconfigerror istio system pod istio citadel prpwn crashloopbackoff istio system pod istio cleanup secrets completed istio system pod istio egressgateway running istio system pod istio galley jscww containercreating istio system pod istio sidecar injector containercreating version istioctl version version gitrevision user root hub gcr io istio release golangversion buildstatus clean kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux is istio auth enabled or not no idea environment kops w coreos on aws cluster state | 0 |
3,023 | 31,623,332,118 | IssuesEvent | 2023-09-06 02:02:13 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | closed | [FEA] Add retry support to GpuSorter.fullySortBatch | feature request reliability | **Is your feature request related to a problem? Please describe.**
This is used both for GpuTopN and GpuSortEachBatchIterator.
| True | [FEA] Add retry support to GpuSorter.fullySortBatch - **Is your feature request related to a problem? Please describe.**
This is used both for GpuTopN and GpuSortEachBatchIterator.
| reli | add retry support to gpusorter fullysortbatch is your feature request related to a problem please describe this is used both for gputopn and gpusorteachbatchiterator | 1 |
2,238 | 24,364,117,445 | IssuesEvent | 2022-10-03 13:58:39 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | closed | [FEA] Allow broadcast tables to be spilled if necessary | P1 reliability | **Is your feature request related to a problem? Please describe.**
Currently broadcast tables are "intentionally leaked" in GPU memory, as they are only cleaned up when garbage collected. It would be nice if instead of leaving them permanently in GPU memory until garbage collected we were able to spill them to host memory (and ultimately disk if necessary) when GPU memory is low.
**Describe the solution you'd like**
We should add broadcast tables to the spillable buffer framework. There could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used (as the spill framework currently doesn't migrate a buffer's recorded location from host back to device once spilled). However running a bit slow beats crashing due to OOM, so this would be a good first step. | True | [FEA] Allow broadcast tables to be spilled if necessary - **Is your feature request related to a problem? Please describe.**
Currently broadcast tables are "intentionally leaked" in GPU memory, as they are only cleaned up when garbage collected. It would be nice if instead of leaving them permanently in GPU memory until garbage collected we were able to spill them to host memory (and ultimately disk if necessary) when GPU memory is low.
**Describe the solution you'd like**
We should add broadcast tables to the spillable buffer framework. There could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used (as the spill framework currently doesn't migrate a buffer's recorded location from host back to device once spilled). However running a bit slow beats crashing due to OOM, so this would be a good first step. | reli | allow broadcast tables to be spilled if necessary is your feature request related to a problem please describe currently broadcast tables are intentionally leaked in gpu memory as they are only cleaned up when garbage collected it would be nice if instead of leaving them permanently in gpu memory until garbage collected we were able to spill them to host memory and ultimately disk if necessary when gpu memory is low describe the solution you d like we should add broadcast tables to the spillable buffer framework there could be some performance considerations if broadcast tables are spilled and constantly fetched from host memory to be used as the spill framework currently doesn t migrate a buffer s recorded location from host back to device once spilled however running a bit slow beats crashing due to oom so this would be a good first step | 1 |
209,550 | 16,039,599,602 | IssuesEvent | 2021-04-22 05:46:22 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | opened | [CI] AutoFollowIT testDataStreams_autoFollowAfterDataStreamCreated failing | :Core/Features/Data streams :Distributed/CCR >test-failure | Subsequent tests failed as well, but I suspect they are due to this initial failure.
```
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400}
```
**Build scan:**
https://gradle-enterprise.elastic.co/s/q4m7uwwqj7mei/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testDataStreams_autoFollowAfterDataStreamCreated
**Reproduction line:**
`./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated" -Dtests.seed=197AA5A92FE81C06 -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Argentina/San_Juan -Druntime.java=8`
**Applicable branches:**
7.x
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testDataStreams_autoFollowAfterDataStreamCreated
**Failure excerpt:**
```
org.elasticsearch.client.ResponseException: method [PUT], host [http://[::1]:45566], URI [/.ds-logs-syslog-prod-2021.04.22-000001/_ccr/follow?wait_for_active_shards=1], status line [HTTP/1.1 400 Bad Request]
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400}
at __randomizedtesting.SeedInfo.seed([197AA5A92FE81C06:AD357CF326BDC165]:0)
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:109)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:81)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:77)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:73)
at org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated(AutoFollowIT.java:319)
at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
``` | 1.0 | [CI] AutoFollowIT testDataStreams_autoFollowAfterDataStreamCreated failing - Subsequent tests failed as well, but I suspect they are due to this initial failure.
```
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400}
```
**Build scan:**
https://gradle-enterprise.elastic.co/s/q4m7uwwqj7mei/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testDataStreams_autoFollowAfterDataStreamCreated
**Reproduction line:**
`./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated" -Dtests.seed=197AA5A92FE81C06 -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Argentina/San_Juan -Druntime.java=8`
**Applicable branches:**
7.x
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testDataStreams_autoFollowAfterDataStreamCreated
**Failure excerpt:**
```
org.elasticsearch.client.ResponseException: method [PUT], host [http://[::1]:45566], URI [/.ds-logs-syslog-prod-2021.04.22-000001/_ccr/follow?wait_for_active_shards=1], status line [HTTP/1.1 400 Bad Request]
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"}],"type":"illegal_argument_exception","reason":"the leader index settings [{\"index.lifecycle.indexing_complete\":\"true\",\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] and follower index settings [{\"index.lifecycle.name\":\"logs\",\"index.number_of_shards\":\"1\"}] must be identical"},"status":400}
at __randomizedtesting.SeedInfo.seed([197AA5A92FE81C06:AD357CF326BDC165]:0)
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:109)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:81)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:77)
at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.followIndex(ESCCRRestTestCase.java:73)
at org.elasticsearch.xpack.ccr.AutoFollowIT.testDataStreams_autoFollowAfterDataStreamCreated(AutoFollowIT.java:319)
at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
``` | non_reli | autofollowit testdatastreams autofollowafterdatastreamcreated failing subsequent tests failed as well but i suspect they are due to this initial failure error root cause and follower index settings must be identical type illegal argument exception reason the leader index settings and follower index settings must be identical status build scan reproduction line gradlew x pack plugin ccr qa multi cluster follow cluster tests org elasticsearch xpack ccr autofollowit testdatastreams autofollowafterdatastreamcreated dtests seed dtests locale sr latn ba dtests timezone america argentina san juan druntime java applicable branches x reproduces locally no failure history failure excerpt org elasticsearch client responseexception method host uri status line error root cause and follower index settings must be identical type illegal argument exception reason the leader index settings and follower index settings must be identical status at randomizedtesting seedinfo seed at org elasticsearch client restclient convertresponse restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr esccrresttestcase followindex esccrresttestcase java at org elasticsearch xpack ccr autofollowit testdatastreams autofollowafterdatastreamcreated autofollowit java at sun reflect nativemethodaccessorimpl nativemethodaccessorimpl java at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at java lang thread run thread java | 0 |
199,924 | 6,996,156,657 | IssuesEvent | 2017-12-15 22:41:47 | GoogleCloudPlatform/google-cloud-go | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-go | closed | logging: automatically detect GKE and GAE Flex resources | api: logging help wanted priority: p2 type: feature request | The creation of a new Logger can automatically set ressources for GCE instances, but it cannot do it yet for GKE Containers. Having type:gce_instance can be confusing for code running inside a container. You end up with logs from the containers appear as if the hosting machine was producing them.
Proposed behaviour:
* When code runs inside a Google Container Engine
* the type is set to 'container',
* the labels: pod_id, cluster_name, container_name, namespace_id, are populated as well.
| 1.0 | logging: automatically detect GKE and GAE Flex resources - The creation of a new Logger can automatically set ressources for GCE instances, but it cannot do it yet for GKE Containers. Having type:gce_instance can be confusing for code running inside a container. You end up with logs from the containers appear as if the hosting machine was producing them.
Proposed behaviour:
* When code runs inside a Google Container Engine
* the type is set to 'container',
* the labels: pod_id, cluster_name, container_name, namespace_id, are populated as well.
| non_reli | logging automatically detect gke and gae flex resources the creation of a new logger can automatically set ressources for gce instances but it cannot do it yet for gke containers having type gce instance can be confusing for code running inside a container you end up with logs from the containers appear as if the hosting machine was producing them proposed behaviour when code runs inside a google container engine the type is set to container the labels pod id cluster name container name namespace id are populated as well | 0 |
1,598 | 17,376,274,704 | IssuesEvent | 2021-07-30 21:56:58 | clowdr-app/clowdr | https://api.github.com/repos/clowdr-app/clowdr | closed | Automatically clean up phantom room participants | A-rooms C-bug D-medium S-actions T-reliability U-soon | If you join the room, there is none there, but they are still listed as active
<img width="494" alt="Screen Shot 2021-01-21 at 5 31 26 PM" src="https://user-images.githubusercontent.com/506426/105420581-9f7d0400-5c0e-11eb-8e36-07e627d21c63.png">
| True | Automatically clean up phantom room participants - If you join the room, there is none there, but they are still listed as active
<img width="494" alt="Screen Shot 2021-01-21 at 5 31 26 PM" src="https://user-images.githubusercontent.com/506426/105420581-9f7d0400-5c0e-11eb-8e36-07e627d21c63.png">
| reli | automatically clean up phantom room participants if you join the room there is none there but they are still listed as active img width alt screen shot at pm src | 1 |
401 | 7,308,657,544 | IssuesEvent | 2018-02-28 09:10:36 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Crash in SqlConnection.Close() when using TransactionScope on linux. | area-System.Data.SqlClient os-linux tenet-reliability | I'm using SqlClient 4.5.0-preview1-26207-04 in netcoreapp2.0.
```
.NET Command Line Tools (2.0.0)
Product Information:
Version: 2.0.0
Commit SHA-1 hash: cdcd1928c9
Runtime Environment:
OS Name: centos
OS Version: 7
OS Platform: Linux
RID: centos.7-x64
Base Path: /usr/share/dotnet/sdk/2.0.0/
Microsoft .NET Core Shared Framework Host
Version : 2.0.0
Build : e8b8861ac7faf042c87a5c2f9f2d04c98b69f28d
```
The following snippet would crash on the above machine (removed some code for brevity). Works without issues on Windows.
```c#
using (var txscope = new TransactionScope(TransactionScopeAsyncFlowOption.Enabled))
using (var cn = _dBProvider.GetWriteConnection())
{
await cn.OpenAsync();
if (cmd.RecordCertificationStatus)
{
await cn.ExecuteAsync(@"INSERT INTO (...) (@iduser, @idtitle, @value, getdate()); ",
new { @iduser = cmd.SubId, @idtitle=cmd.EntityId, @value = cmd.Model.IsCertified });
}
await cn.ExecuteAsync(sql, data);
txscope.Complete();
}
```
Both statements are inserts. Stacktrace is below:
```
"System.NullReferenceException: Object reference not set to an instance of an object.
at System.Data.ProviderBase.DbConnectionPool.DeactivateObject(DbConnectionInternal obj)
at System.Data.ProviderBase.DbConnectionPool.PutObject(DbConnectionInternal obj, Object owningObject)
at System.Data.ProviderBase.DbConnectionInternal.CloseConnection(DbConnection owningObject, DbConnectionFactory connectionFactory)
at System.Data.SqlClient.SqlConnection.Close()
at System.Data.SqlClient.SqlConnection.Dispose(Boolean disposing)
at System.ComponentModel.Component.Dispose()
at ...TitleUpdate.Handler.<HandleCore>d__2.MoveNext() in ...TitleUpdate.cs:line 189
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at MediatR.Internal.RequestHandlerWrapperImpl`1.<>c__DisplayClass0_0.<<Handle>g__Handler0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
at ...ValidationBehavior`2.<Handle>d__2.MoveNext() in ...ValidationBehavior.cs:line 34
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at ....TitlesController.<Update>d__5.MoveNext() in TitleController.cs:line 98
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeActionMethodAsync>d__12.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeNextActionFilterAsync>d__10.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Rethrow(ActionExecutedContext context)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Next(State& next, Scope& scope, Object& state, Boolean& isCompleted)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeInnerFilterAsync>d__14.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ResourceInvoker.<InvokeNextExceptionFilterAsync>d__23.MoveNext()"
``` | True | Crash in SqlConnection.Close() when using TransactionScope on linux. - I'm using SqlClient 4.5.0-preview1-26207-04 in netcoreapp2.0.
```
.NET Command Line Tools (2.0.0)
Product Information:
Version: 2.0.0
Commit SHA-1 hash: cdcd1928c9
Runtime Environment:
OS Name: centos
OS Version: 7
OS Platform: Linux
RID: centos.7-x64
Base Path: /usr/share/dotnet/sdk/2.0.0/
Microsoft .NET Core Shared Framework Host
Version : 2.0.0
Build : e8b8861ac7faf042c87a5c2f9f2d04c98b69f28d
```
The following snippet would crash on the above machine (removed some code for brevity). Works without issues on Windows.
```c#
using (var txscope = new TransactionScope(TransactionScopeAsyncFlowOption.Enabled))
using (var cn = _dBProvider.GetWriteConnection())
{
await cn.OpenAsync();
if (cmd.RecordCertificationStatus)
{
await cn.ExecuteAsync(@"INSERT INTO (...) (@iduser, @idtitle, @value, getdate()); ",
new { @iduser = cmd.SubId, @idtitle=cmd.EntityId, @value = cmd.Model.IsCertified });
}
await cn.ExecuteAsync(sql, data);
txscope.Complete();
}
```
Both statements are inserts. Stacktrace is below:
```
"System.NullReferenceException: Object reference not set to an instance of an object.
at System.Data.ProviderBase.DbConnectionPool.DeactivateObject(DbConnectionInternal obj)
at System.Data.ProviderBase.DbConnectionPool.PutObject(DbConnectionInternal obj, Object owningObject)
at System.Data.ProviderBase.DbConnectionInternal.CloseConnection(DbConnection owningObject, DbConnectionFactory connectionFactory)
at System.Data.SqlClient.SqlConnection.Close()
at System.Data.SqlClient.SqlConnection.Dispose(Boolean disposing)
at System.ComponentModel.Component.Dispose()
at ...TitleUpdate.Handler.<HandleCore>d__2.MoveNext() in ...TitleUpdate.cs:line 189
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at MediatR.Internal.RequestHandlerWrapperImpl`1.<>c__DisplayClass0_0.<<Handle>g__Handler0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
at ...ValidationBehavior`2.<Handle>d__2.MoveNext() in ...ValidationBehavior.cs:line 34
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at ....TitlesController.<Update>d__5.MoveNext() in TitleController.cs:line 98
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeActionMethodAsync>d__12.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeNextActionFilterAsync>d__10.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Rethrow(ActionExecutedContext context)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.Next(State& next, Scope& scope, Object& state, Boolean& isCompleted)
at Microsoft.AspNetCore.Mvc.Internal.ControllerActionInvoker.<InvokeInnerFilterAsync>d__14.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.AspNetCore.Mvc.Internal.ResourceInvoker.<InvokeNextExceptionFilterAsync>d__23.MoveNext()"
``` | reli | crash in sqlconnection close when using transactionscope on linux i m using sqlclient in net command line tools product information version commit sha hash runtime environment os name centos os version os platform linux rid centos base path usr share dotnet sdk microsoft net core shared framework host version build the following snippet would crash on the above machine removed some code for brevity works without issues on windows c using var txscope new transactionscope transactionscopeasyncflowoption enabled using var cn dbprovider getwriteconnection await cn openasync if cmd recordcertificationstatus await cn executeasync insert into iduser idtitle value getdate new iduser cmd subid idtitle cmd entityid value cmd model iscertified await cn executeasync sql data txscope complete both statements are inserts stacktrace is below system nullreferenceexception object reference not set to an instance of an object at system data providerbase dbconnectionpool deactivateobject dbconnectioninternal obj at system data providerbase dbconnectionpool putobject dbconnectioninternal obj object owningobject at system data providerbase dbconnectioninternal closeconnection dbconnection owningobject dbconnectionfactory connectionfactory at system data sqlclient sqlconnection close at system data sqlclient sqlconnection dispose boolean disposing at system componentmodel component dispose at titleupdate handler d movenext in titleupdate cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at mediatr internal requesthandlerwrapperimpl c g d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at validationbehavior d movenext in validationbehavior cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at titlescontroller d movenext in titlecontroller cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft aspnetcore mvc internal controlleractioninvoker rethrow actionexecutedcontext context at microsoft aspnetcore mvc internal controlleractioninvoker next state next scope scope object state boolean iscompleted at microsoft aspnetcore mvc internal controlleractioninvoker d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft aspnetcore mvc internal resourceinvoker d movenext | 1 |
383,407 | 11,355,798,045 | IssuesEvent | 2020-01-24 20:56:43 | yalla-coop/presspad | https://api.github.com/repos/yalla-coop/presspad | opened | Create updates for intern, hosts, and organisations | priority-5 | - [ ] When intern request a stay
- [ ] host (`stayRequest`)
- [ ] org (`stayRequest`)
- [ ] when intern accept a request
- [ ] intern (`stayApproved`)
- [ ] org (`stayApproved`)
- [ ] when host reject a request
- [ ] intern (`stayRejected`)
- [ ] org (`stayRejected`)
- [ ] when stay completed
- [ ] intern (`giveReviewReminder`)
- [ ] host (`giveReviewReminder`)
- [ ] org (`stayCompleted`)
- [ ] when host give an intern a review
- [ ] intern (`getReview`)
- [ ] org (`getReview`)
- [ ] when intern give a host a review
- [ ] host (`getReview`)
- [ ] org (`internGaveHostReview`)
| 1.0 | Create updates for intern, hosts, and organisations - - [ ] When intern request a stay
- [ ] host (`stayRequest`)
- [ ] org (`stayRequest`)
- [ ] when intern accept a request
- [ ] intern (`stayApproved`)
- [ ] org (`stayApproved`)
- [ ] when host reject a request
- [ ] intern (`stayRejected`)
- [ ] org (`stayRejected`)
- [ ] when stay completed
- [ ] intern (`giveReviewReminder`)
- [ ] host (`giveReviewReminder`)
- [ ] org (`stayCompleted`)
- [ ] when host give an intern a review
- [ ] intern (`getReview`)
- [ ] org (`getReview`)
- [ ] when intern give a host a review
- [ ] host (`getReview`)
- [ ] org (`internGaveHostReview`)
| non_reli | create updates for intern hosts and organisations when intern request a stay host stayrequest org stayrequest when intern accept a request intern stayapproved org stayapproved when host reject a request intern stayrejected org stayrejected when stay completed intern givereviewreminder host givereviewreminder org staycompleted when host give an intern a review intern getreview org getreview when intern give a host a review host getreview org interngavehostreview | 0 |
35,742 | 5,005,664,868 | IssuesEvent | 2016-12-12 11:28:49 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | [Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster | kind/upgrade-test-failure non-release-blocker | Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4)
Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
| 1.0 | [Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster - Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4)
Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
| non_reli | scheduledjob should replace jobs when replaceconcurrent in ci kubernetes gce upgrade cluster test has been failing in sample failure based on spreadsheet tracking upgrade test failures created by krousey | 0 |
152,864 | 19,697,885,119 | IssuesEvent | 2022-01-12 14:00:30 | rarkins/spdx-maven-plugin-demo | https://api.github.com/repos/rarkins/spdx-maven-plugin-demo | opened | CVE-2020-35490 (High) detected in jackson-databind-2.9.10.4.jar | security vulnerability | ## CVE-2020-35490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p>
<p>
Dependency Hierarchy:
- spdx-tools-2.2.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-35490 (High) detected in jackson-databind-2.9.10.4.jar - ## CVE-2020-35490 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p>
<p>
Dependency Hierarchy:
- spdx-tools-2.2.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_reli | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spdx tools jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
1,972 | 22,296,923,997 | IssuesEvent | 2022-06-13 03:27:02 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | closed | [FEA] Stop running task attempts on executors that encounter "sticky" CUDA errors | P1 reliability | **Is your feature request related to a problem? Please describe.**
Certain CUDA errors, like illegal memory access, are "sticky," meaning that all CUDA operations to the GPU after the error will continue to return the same error over and over. No GPU operations will succeed after that point.
**Describe the solution you'd like**
The RAPIDS Accelerator should take measures to prevent further task execution on the executor once these "sticky" exceptions are detected. Tearing down the executor process is probably the best option, at least in the short-term. Without an external shuffle handler we will lose the shuffle of tasks that have completed, but this is probably a better way to "fail fast" then allow the executor to keep accepting new tasks only to have them fail the first time they touch the GPU.
| True | [FEA] Stop running task attempts on executors that encounter "sticky" CUDA errors - **Is your feature request related to a problem? Please describe.**
Certain CUDA errors, like illegal memory access, are "sticky," meaning that all CUDA operations to the GPU after the error will continue to return the same error over and over. No GPU operations will succeed after that point.
**Describe the solution you'd like**
The RAPIDS Accelerator should take measures to prevent further task execution on the executor once these "sticky" exceptions are detected. Tearing down the executor process is probably the best option, at least in the short-term. Without an external shuffle handler we will lose the shuffle of tasks that have completed, but this is probably a better way to "fail fast" then allow the executor to keep accepting new tasks only to have them fail the first time they touch the GPU.
| reli | stop running task attempts on executors that encounter sticky cuda errors is your feature request related to a problem please describe certain cuda errors like illegal memory access are sticky meaning that all cuda operations to the gpu after the error will continue to return the same error over and over no gpu operations will succeed after that point describe the solution you d like the rapids accelerator should take measures to prevent further task execution on the executor once these sticky exceptions are detected tearing down the executor process is probably the best option at least in the short term without an external shuffle handler we will lose the shuffle of tasks that have completed but this is probably a better way to fail fast then allow the executor to keep accepting new tasks only to have them fail the first time they touch the gpu | 1 |
2,229 | 24,319,431,624 | IssuesEvent | 2022-09-30 09:23:48 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | closed | DefaultActorClock is not thread safe | kind/bug area/performance severity/high severity/mid area/reliability | **Describe the bug**
It seems we now create the ActorScheduler with a predefined actor clock, either ControlledClock or the DefaultActorClock https://github.com/camunda/zeebe/blob/main/dist/src/main/java/io/camunda/zeebe/shared/ActorClockConfiguration.java#L34-L35
The problem is that the DefaultActorClock is not thread safe! Previously we created the ActorScheduler without clock, which caused [that each thread has its own clock](https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L64). Then there is no issue.
If we use the DefaultActorClock in all threads this might cause issues on updating the time https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/clock/DefaultActorClock.java#L24-L31 which is not thread safe and called by all threads https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L78
and also when timers are scheduled https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorTimerQueue.java#L53
I found that during investigating https://github.com/camunda/zeebe/pull/10390 and trying to make ActorTimerQueue threadsafe...
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
<!--
Steps to reproduce the behavior
If possible add a minimal reproducer code sample
- when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java
-->
**Expected behavior**
Either we not set the clock or we use a thread safe structure.
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<!-- If possible add the full stacktrace or Zeebe log which contains the issue. -->
<details><summary>Full Stacktrace</summary>
<p>
```
<STACKTRACE>
```
</p>
</details>
**Environment:**
- OS: <!-- [e.g. Linux] -->
- Zeebe Version: 8.x<!-- [e.g. 0.20.0] -->
- Configuration: <!-- [e.g. exporters etc.] -->
| True | DefaultActorClock is not thread safe - **Describe the bug**
It seems we now create the ActorScheduler with a predefined actor clock, either ControlledClock or the DefaultActorClock https://github.com/camunda/zeebe/blob/main/dist/src/main/java/io/camunda/zeebe/shared/ActorClockConfiguration.java#L34-L35
The problem is that the DefaultActorClock is not thread safe! Previously we created the ActorScheduler without clock, which caused [that each thread has its own clock](https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L64). Then there is no issue.
If we use the DefaultActorClock in all threads this might cause issues on updating the time https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/clock/DefaultActorClock.java#L24-L31 which is not thread safe and called by all threads https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorThread.java#L78
and also when timers are scheduled https://github.com/camunda/zeebe/blob/main/scheduler/src/main/java/io/camunda/zeebe/scheduler/ActorTimerQueue.java#L53
I found that during investigating https://github.com/camunda/zeebe/pull/10390 and trying to make ActorTimerQueue threadsafe...
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
<!--
Steps to reproduce the behavior
If possible add a minimal reproducer code sample
- when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java
-->
**Expected behavior**
Either we not set the clock or we use a thread safe structure.
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<!-- If possible add the full stacktrace or Zeebe log which contains the issue. -->
<details><summary>Full Stacktrace</summary>
<p>
```
<STACKTRACE>
```
</p>
</details>
**Environment:**
- OS: <!-- [e.g. Linux] -->
- Zeebe Version: 8.x<!-- [e.g. 0.20.0] -->
- Configuration: <!-- [e.g. exporters etc.] -->
| reli | defaultactorclock is not thread safe describe the bug it seems we now create the actorscheduler with a predefined actor clock either controlledclock or the defaultactorclock the problem is that the defaultactorclock is not thread safe previously we created the actorscheduler without clock which caused then there is no issue if we use the defaultactorclock in all threads this might cause issues on updating the time which is not thread safe and called by all threads and also when timers are scheduled i found that during investigating and trying to make actortimerqueue threadsafe to reproduce steps to reproduce the behavior if possible add a minimal reproducer code sample when using the java client expected behavior either we not set the clock or we use a thread safe structure log stacktrace full stacktrace environment os zeebe version x configuration | 1 |
687,162 | 23,515,782,898 | IssuesEvent | 2022-08-18 21:11:51 | ezolenko/rollup-plugin-typescript2 | https://api.github.com/repos/ezolenko/rollup-plugin-typescript2 | closed | some type-only TS files are ignored when using `tsconfig` `files`? | kind: bug solution: duplicate problem: removed issue template priority: in progress problem: plugin order topic: type-only / emit-less imports | ## What happens and why it is incorrect
there are many files ending in `*.ts` in the src, but only some have been transpiled
1. Does `tsc` have the same output? If so, please explain why this is incorrect behavior
NO, `tsc` run as expectly.

here's file structure:

but rpt2 seems to ignore the file named `component.ts`:
<img width="1241" alt="image" src="https://user-images.githubusercontent.com/52886395/185072856-019e3e3e-9598-4a5e-8dfb-10cb9f7276cc.png">
## Environment
### Versions
```text
System:
OS: macOS 11.4
CPU: (12) x64 Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz
Memory: 59.11 MB / 32.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 14.17.4 - ~/.nvm/versions/node/v14.17.4/bin/node
Yarn: 1.22.11 - ~/.nvm/versions/node/v14.17.4/bin/yarn
npm: 6.14.14 - ~/.nvm/versions/node/v14.17.4/bin/npm
npmPackages:
rollup: ^2.45.2 => 2.45.2
rollup-plugin-typescript2: ^0.30.0 => 0.30.0
typescript: ^4.3.5 => 4.3.5
npmGlobalPackages:
typescript: 4.3.5
```
<!--- paste your rollup config below if relevant --->
<details>
<summary><h4><code>rollup.config.js</code></h4>: </summary>
<!--- INSERT rollup.config.ts IN THE CODE SNIPPET BELOW --->
```js
import { nodeResolve } from '@rollup/plugin-node-resolve';
import replace from '@rollup/plugin-replace';
import postcss from 'rollup-plugin-postcss';
import image from '@rollup/plugin-image';
import ts from 'rollup-plugin-typescript2';
import dts from 'rollup-plugin-dts';
import json from '@rollup/plugin-json';
import alias from '@rollup/plugin-alias';
import autoExternal from 'rollup-plugin-auto-external';
import path from 'path';
import importCss from './rollup-plugin-import-css';
const resolve = (...dirs) => path.resolve(__dirname, '../', ...dirs);
function toUpperCase(match) {
return match.replace('-', '').toUpperCase();
}
export default [
{
input: resolve('src/lib/index.tsx'),
output: [
{
file: resolve('dist/index.esm.js'),
format: 'es',
sourcemap: true,
},
],
external: id => {
try {
const idSourcePath = require.resolve(id, { paths: [resolve()] });
return idSourcePath && idSourcePath.includes('node_modules');
} catch (error) {
return false;
}
},
plugins: [
ts({
check: false,
tsconfig: resolve('tsconfig.json'),
tsconfigOverride: {
compilerOptions: { declaration: true, declarationDir: resolve('dist/type') },
},
verbosity: 2,
useTsconfigDeclarationDir: true,
include: ['*.ts+(|x)', '**/*.ts+(|x)', '*.js+(|x)', '**/*.js+(|x)'],
}),
image(),
postcss({
extensions: ['.css', '.scss', '.less'],
autoModules: true,
extract: 'index.css',
namedExports(name) {
let reg = /-[a-z]/g;
const temp = name.replace(reg, toUpperCase);
return temp;
},
}),
json(),
alias({
entries: [
{
find: 'src',
replacement: resolve('src'),
},
],
}),
nodeResolve({
extensions: ['.js', '.jsx', '.ts', '.tsx'],
// some package.json files have a "browser" field which specifies
// alternative files to load for people bundling for the browser. If
// that's you, either use this option or add "browser" to the
// "mainfields" option, otherwise pkg.browser will be ignored
browser: true,
preferBuiltins: true,
mainFields: ['browser', 'jsnext', 'module', 'main'],
}),
replace({
'process.env.NODE_ENV': JSON.stringify('development'),
'process.env.SEMI_ICON_LAZY_LOAD': true,
preventAssignment: true,
}),
autoExternal({
packagePath: resolve(),
}),
importCss(),
],
},
{
input: resolve('dist/type/src/lib/index.d.ts'),
external: id => {
try {
const idSourcePath = require.resolve(id, { paths: [resolve()] });
return idSourcePath && idSourcePath.includes('node_modules');
} catch (error) {
return false;
}
},
output: [{ file: resolve('dist/index.d.ts'), format: 'es' }],
plugins: [dts()],
},
];
```
</details>
<!--- paste your tsconfig.json below if relevant --->
<details>
<summary><h4><code>tsconfig.json</code></h4>: </summary>
<!--- INSERT tsconfig.json IN THE CODE SNIPPET BELOW --->
```json5
{
"compilerOptions": {
"target": "es6",
"module": "esnext",
"lib": ["es7", "dom"],
"sourceMap": true,
"allowJs": true,
"jsx": "react",
"moduleResolution": "node",
"experimentalDecorators": true,
"rootDir": "./",
"baseUrl": "./src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"noEmit": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-editor": ["src/lib/*"]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
]
},
"typeRoots": ["node", "node_modules/@types", "./src/typings"],
"exclude": ["node_modules"]
}
```
</details>
<!--- paste your package.json below if relevant --->
<details>
<summary><h4><code>package.json</code></h4>: </summary>
<!--- INSERT package.json IN THE CODE SNIPPET BELOW --->
```json
```
</details>
<!--- add verbosity verbosity: 3 to plugin options and attach output if relevant (censor out anything sensitive) --->
<details>
<summary><h4>plugin output with verbosity 3</h4>: </summary>
<!--- INSERT plugin output IN THE CODE SNIPPET BELOW or attach --->
```text
/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx → dist/index.esm.js...
rpt2: built-in options overrides: {
"noEmitHelpers": false,
"importHelpers": true,
"noResolve": false,
"noEmit": false,
"inlineSourceMap": false,
"outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder",
"moduleResolution": 2,
"allowNonTsExtensions": true
}
rpt2: parsed tsconfig: {
"options": {
"isolatedModules": false,
"declaration": true,
"target": 2,
"module": 99,
"lib": [
"lib.es2016.d.ts",
"lib.dom.d.ts"
],
"sourceMap": true,
"allowJs": true,
"jsx": 2,
"moduleResolution": 2,
"experimentalDecorators": true,
"rootDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"baseUrl": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-tag-table": [
"src/lib/*"
]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
],
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type",
"configFilePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json",
"pathsBasePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"noEmitHelpers": false,
"noResolve": false,
"noEmit": false,
"inlineSourceMap": false,
"outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder",
"allowNonTsExtensions": true
},
"fileNames": [
"/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx"
],
"typeAcquisition": {
"enable": false,
"include": [],
"exclude": []
},
"raw": {
"files": [
"src/lib/index.tsx"
],
"compilerOptions": {
"isolatedModules": false,
"declaration": true,
"target": "es6",
"module": "esnext",
"lib": [
"es7",
"dom"
],
"sourceMap": true,
"allowJs": true,
"jsx": "react",
"moduleResolution": "node",
"experimentalDecorators": true,
"rootDir": "./",
"baseUrl": "./src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-tag-table": [
"src/lib/*"
]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
],
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type"
},
"typeRoots": [
"node",
"node_modules/@types",
"./src/typings"
],
"exclude": [
"node_modules"
],
"compileOnSave": false
},
"errors": [],
"wildcardDirectories": {},
"compileOnSave": false
}
rpt2: typescript version: 4.3.5
rpt2: tslib version: 2.1.0
rpt2: rollup version: 2.45.2
rpt2: rollup-plugin-typescript2 version: 0.30.0
rpt2: plugin options:
{
"check": false,
"tsconfig": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json",
"tsconfigOverride": {
"compilerOptions": {
"declaration": true,
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type"
}
},
"verbosity": 3,
"useTsconfigDeclarationDir": true,
"include": [
"*.ts+(|x)",
"**/*.ts+(|x)",
"*.js+(|x)",
"**/*.js+(|x)"
],
"clean": false,
"cacheRoot": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2",
"exclude": [
"*.d.ts",
"**/*.d.ts"
],
"abortOnError": true,
"rollupCommonJSResolveHack": false,
"transformers": [],
"tsconfigDefaults": {},
"objectHashIgnoreUnknownHack": false,
"cwd": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"typescript": "version 4.3.5"
}
rpt2: rollup config:
{
"input": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx",
"plugins": [
{
"name": "rpt2"
},
{
"name": "image"
},
{
"name": "postcss"
},
{
"name": "json"
},
{
"name": "alias"
},
{
"name": "node-resolve"
},
{
"name": "replace"
},
{
"name": "auto-external"
},
{},
{
"name": "stdin"
}
],
"output": [
{
"file": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/index.esm.js",
"format": "es",
"plugins": [],
"sourcemap": true
}
]
}
rpt2: tsconfig path: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json
rpt2: included:
[
"*.ts+(|x)",
"**/*.ts+(|x)",
"*.js+(|x)",
"**/*.js+(|x)"
]
rpt2: excluded:
[
"*.d.ts",
"**/*.d.ts"
]
rpt2: Ambient types:
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__core/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__generator/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__template/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__traverse/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/color-name/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/css-modules-loader-core/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/estree/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/graceful-fs/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/hast/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-coverage/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-report/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-reports/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/jest/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/mdast/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse5/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prettier/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prop-types/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/q/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react-dom/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/resolve/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/stack-utils/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/unist/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs-parser/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts
rpt2: /Users/bytedance/node_modules/@types/json5/index.d.ts
rpt2: /Users/bytedance/node_modules/@types/minimist/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts
rpt2: ambient types changed, redoing all semantic diagnostics
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/7cdbf1c7db69c2fd53815824bca49513a129c642'
rpt2: cache miss
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: resolving './utils' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: resolving './mock' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/a3be5c51ebb3fef03fc9a43c539d188ad1577216'
rpt2: cache hit
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/d1d6e9acec5e9e0b4392fcbc5bf623dab081aa31'
rpt2: cache hit
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: generating target 1
rpt2: rolling caches
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/index.d.ts'
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/mock.d.ts'
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/utils.d.ts'
(!) Broken sourcemap
https://rollupjs.org/guide/en/#warning-sourcemap-is-likely-to-be-incorrect
Plugins that transform code (such as 'at position 9') should generate accompanying sourcemaps
created dist/index.esm.js in 2.7s
User/dist/type/src/lib/index.d.ts → dist/index.d.ts...
[!] Error: Could not resolve './component' from dist/type/src/lib/index.d.ts
Error: Could not resolve './component' from dist/type/src/lib/index.d.ts
```
</details>
| 1.0 | some type-only TS files are ignored when using `tsconfig` `files`? - ## What happens and why it is incorrect
there are many files ending in `*.ts` in the src, but only some have been transpiled
1. Does `tsc` have the same output? If so, please explain why this is incorrect behavior
NO, `tsc` run as expectly.

here's file structure:

but rpt2 seems to ignore the file named `component.ts`:
<img width="1241" alt="image" src="https://user-images.githubusercontent.com/52886395/185072856-019e3e3e-9598-4a5e-8dfb-10cb9f7276cc.png">
## Environment
### Versions
```text
System:
OS: macOS 11.4
CPU: (12) x64 Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz
Memory: 59.11 MB / 32.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 14.17.4 - ~/.nvm/versions/node/v14.17.4/bin/node
Yarn: 1.22.11 - ~/.nvm/versions/node/v14.17.4/bin/yarn
npm: 6.14.14 - ~/.nvm/versions/node/v14.17.4/bin/npm
npmPackages:
rollup: ^2.45.2 => 2.45.2
rollup-plugin-typescript2: ^0.30.0 => 0.30.0
typescript: ^4.3.5 => 4.3.5
npmGlobalPackages:
typescript: 4.3.5
```
<!--- paste your rollup config below if relevant --->
<details>
<summary><h4><code>rollup.config.js</code></h4>: </summary>
<!--- INSERT rollup.config.ts IN THE CODE SNIPPET BELOW --->
```js
import { nodeResolve } from '@rollup/plugin-node-resolve';
import replace from '@rollup/plugin-replace';
import postcss from 'rollup-plugin-postcss';
import image from '@rollup/plugin-image';
import ts from 'rollup-plugin-typescript2';
import dts from 'rollup-plugin-dts';
import json from '@rollup/plugin-json';
import alias from '@rollup/plugin-alias';
import autoExternal from 'rollup-plugin-auto-external';
import path from 'path';
import importCss from './rollup-plugin-import-css';
const resolve = (...dirs) => path.resolve(__dirname, '../', ...dirs);
function toUpperCase(match) {
return match.replace('-', '').toUpperCase();
}
export default [
{
input: resolve('src/lib/index.tsx'),
output: [
{
file: resolve('dist/index.esm.js'),
format: 'es',
sourcemap: true,
},
],
external: id => {
try {
const idSourcePath = require.resolve(id, { paths: [resolve()] });
return idSourcePath && idSourcePath.includes('node_modules');
} catch (error) {
return false;
}
},
plugins: [
ts({
check: false,
tsconfig: resolve('tsconfig.json'),
tsconfigOverride: {
compilerOptions: { declaration: true, declarationDir: resolve('dist/type') },
},
verbosity: 2,
useTsconfigDeclarationDir: true,
include: ['*.ts+(|x)', '**/*.ts+(|x)', '*.js+(|x)', '**/*.js+(|x)'],
}),
image(),
postcss({
extensions: ['.css', '.scss', '.less'],
autoModules: true,
extract: 'index.css',
namedExports(name) {
let reg = /-[a-z]/g;
const temp = name.replace(reg, toUpperCase);
return temp;
},
}),
json(),
alias({
entries: [
{
find: 'src',
replacement: resolve('src'),
},
],
}),
nodeResolve({
extensions: ['.js', '.jsx', '.ts', '.tsx'],
// some package.json files have a "browser" field which specifies
// alternative files to load for people bundling for the browser. If
// that's you, either use this option or add "browser" to the
// "mainfields" option, otherwise pkg.browser will be ignored
browser: true,
preferBuiltins: true,
mainFields: ['browser', 'jsnext', 'module', 'main'],
}),
replace({
'process.env.NODE_ENV': JSON.stringify('development'),
'process.env.SEMI_ICON_LAZY_LOAD': true,
preventAssignment: true,
}),
autoExternal({
packagePath: resolve(),
}),
importCss(),
],
},
{
input: resolve('dist/type/src/lib/index.d.ts'),
external: id => {
try {
const idSourcePath = require.resolve(id, { paths: [resolve()] });
return idSourcePath && idSourcePath.includes('node_modules');
} catch (error) {
return false;
}
},
output: [{ file: resolve('dist/index.d.ts'), format: 'es' }],
plugins: [dts()],
},
];
```
</details>
<!--- paste your tsconfig.json below if relevant --->
<details>
<summary><h4><code>tsconfig.json</code></h4>: </summary>
<!--- INSERT tsconfig.json IN THE CODE SNIPPET BELOW --->
```json5
{
"compilerOptions": {
"target": "es6",
"module": "esnext",
"lib": ["es7", "dom"],
"sourceMap": true,
"allowJs": true,
"jsx": "react",
"moduleResolution": "node",
"experimentalDecorators": true,
"rootDir": "./",
"baseUrl": "./src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"noEmit": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-editor": ["src/lib/*"]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
]
},
"typeRoots": ["node", "node_modules/@types", "./src/typings"],
"exclude": ["node_modules"]
}
```
</details>
<!--- paste your package.json below if relevant --->
<details>
<summary><h4><code>package.json</code></h4>: </summary>
<!--- INSERT package.json IN THE CODE SNIPPET BELOW --->
```json
```
</details>
<!--- add verbosity verbosity: 3 to plugin options and attach output if relevant (censor out anything sensitive) --->
<details>
<summary><h4>plugin output with verbosity 3</h4>: </summary>
<!--- INSERT plugin output IN THE CODE SNIPPET BELOW or attach --->
```text
/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx → dist/index.esm.js...
rpt2: built-in options overrides: {
"noEmitHelpers": false,
"importHelpers": true,
"noResolve": false,
"noEmit": false,
"inlineSourceMap": false,
"outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder",
"moduleResolution": 2,
"allowNonTsExtensions": true
}
rpt2: parsed tsconfig: {
"options": {
"isolatedModules": false,
"declaration": true,
"target": 2,
"module": 99,
"lib": [
"lib.es2016.d.ts",
"lib.dom.d.ts"
],
"sourceMap": true,
"allowJs": true,
"jsx": 2,
"moduleResolution": 2,
"experimentalDecorators": true,
"rootDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"baseUrl": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-tag-table": [
"src/lib/*"
]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
],
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type",
"configFilePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json",
"pathsBasePath": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"noEmitHelpers": false,
"noResolve": false,
"noEmit": false,
"inlineSourceMap": false,
"outDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/placeholder",
"allowNonTsExtensions": true
},
"fileNames": [
"/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx"
],
"typeAcquisition": {
"enable": false,
"include": [],
"exclude": []
},
"raw": {
"files": [
"src/lib/index.tsx"
],
"compilerOptions": {
"isolatedModules": false,
"declaration": true,
"target": "es6",
"module": "esnext",
"lib": [
"es7",
"dom"
],
"sourceMap": true,
"allowJs": true,
"jsx": "react",
"moduleResolution": "node",
"experimentalDecorators": true,
"rootDir": "./",
"baseUrl": "./src",
"forceConsistentCasingInFileNames": true,
"noImplicitReturns": true,
"noImplicitThis": false,
"noImplicitAny": false,
"importHelpers": true,
"strictNullChecks": false,
"suppressImplicitAnyIndexErrors": true,
"noUnusedLocals": true,
"allowSyntheticDefaultImports": true,
"esModuleInterop": false,
"paths": {
"@ies/kefu-tag-table": [
"src/lib/*"
]
},
"plugins": [
{
"transform": "ts-optchain/transform"
}
],
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type"
},
"typeRoots": [
"node",
"node_modules/@types",
"./src/typings"
],
"exclude": [
"node_modules"
],
"compileOnSave": false
},
"errors": [],
"wildcardDirectories": {},
"compileOnSave": false
}
rpt2: typescript version: 4.3.5
rpt2: tslib version: 2.1.0
rpt2: rollup version: 2.45.2
rpt2: rollup-plugin-typescript2 version: 0.30.0
rpt2: plugin options:
{
"check": false,
"tsconfig": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json",
"tsconfigOverride": {
"compilerOptions": {
"declaration": true,
"declarationDir": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type"
}
},
"verbosity": 3,
"useTsconfigDeclarationDir": true,
"include": [
"*.ts+(|x)",
"**/*.ts+(|x)",
"*.js+(|x)",
"**/*.js+(|x)"
],
"clean": false,
"cacheRoot": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2",
"exclude": [
"*.d.ts",
"**/*.d.ts"
],
"abortOnError": true,
"rollupCommonJSResolveHack": false,
"transformers": [],
"tsconfigDefaults": {},
"objectHashIgnoreUnknownHack": false,
"cwd": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable",
"typescript": "version 4.3.5"
}
rpt2: rollup config:
{
"input": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx",
"plugins": [
{
"name": "rpt2"
},
{
"name": "image"
},
{
"name": "postcss"
},
{
"name": "json"
},
{
"name": "alias"
},
{
"name": "node-resolve"
},
{
"name": "replace"
},
{
"name": "auto-external"
},
{},
{
"name": "stdin"
}
],
"output": [
{
"file": "/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/index.esm.js",
"format": "es",
"plugins": [],
"sourcemap": true
}
]
}
rpt2: tsconfig path: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/tsconfig.json
rpt2: included:
[
"*.ts+(|x)",
"**/*.ts+(|x)",
"*.js+(|x)",
"**/*.js+(|x)"
]
rpt2: excluded:
[
"*.d.ts",
"**/*.d.ts"
]
rpt2: Ambient types:
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__core/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__generator/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__template/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/babel__traverse/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/color-name/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/css-modules-loader-core/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/estree/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/graceful-fs/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/hast/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-coverage/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-lib-report/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/istanbul-reports/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/jest/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/mdast/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse5/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prettier/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/prop-types/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/q/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/react-dom/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/resolve/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/stack-utils/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/unist/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/yargs-parser/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/json-schema/index.d.ts
rpt2: /Users/bytedance/node_modules/@types/json5/index.d.ts
rpt2: /Users/bytedance/node_modules/@types/minimist/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/node/ts3.7/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/normalize-package-data/index.d.ts
rpt2: /Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/@types/parse-json/index.d.ts
rpt2: ambient types changed, redoing all semantic diagnostics
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/7cdbf1c7db69c2fd53815824bca49513a129c642'
rpt2: cache miss
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: resolving './utils' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: dependency '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: resolving './mock' imported by '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx'
rpt2: to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/a3be5c51ebb3fef03fc9a43c539d188ad1577216'
rpt2: cache hit
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts'
rpt2: transpiling '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: cache: '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/node_modules/.cache/rollup-plugin-typescript2/rpt2_9e69c65b2d74fee0ad388d42568ea4ce7ba1b27d/code/cache/d1d6e9acec5e9e0b4392fcbc5bf623dab081aa31'
rpt2: cache hit
rpt2: generated declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts'
rpt2: generating target 1
rpt2: rolling caches
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/index.tsx' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/index.d.ts'
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/mock.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/mock.d.ts'
rpt2: emitting declarations for '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/src/lib/utils.ts' to '/Users/bytedance/Public/helpdesk-semi-ui/packages/KefuCascaderTable/dist/type/src/lib/utils.d.ts'
(!) Broken sourcemap
https://rollupjs.org/guide/en/#warning-sourcemap-is-likely-to-be-incorrect
Plugins that transform code (such as 'at position 9') should generate accompanying sourcemaps
created dist/index.esm.js in 2.7s
User/dist/type/src/lib/index.d.ts → dist/index.d.ts...
[!] Error: Could not resolve './component' from dist/type/src/lib/index.d.ts
Error: Could not resolve './component' from dist/type/src/lib/index.d.ts
```
</details>
| non_reli | some type only ts files are ignored when using tsconfig files what happens and why it is incorrect there are many files ending in ts in the src but only some have been transpiled does tsc have the same output if so please explain why this is incorrect behavior no, tsc run as expectly here s file structure but seems to ignore the file named component ts img width alt image src environment versions text system os macos cpu intel r core tm cpu memory mb gb shell bin zsh binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm npmpackages rollup rollup plugin typescript npmglobalpackages typescript rollup config js js import noderesolve from rollup plugin node resolve import replace from rollup plugin replace import postcss from rollup plugin postcss import image from rollup plugin image import ts from rollup plugin import dts from rollup plugin dts import json from rollup plugin json import alias from rollup plugin alias import autoexternal from rollup plugin auto external import path from path import importcss from rollup plugin import css const resolve dirs path resolve dirname dirs function touppercase match return match replace touppercase export default input resolve src lib index tsx output file resolve dist index esm js format es sourcemap true external id try const idsourcepath require resolve id paths return idsourcepath idsourcepath includes node modules catch error return false plugins ts check false tsconfig resolve tsconfig json tsconfigoverride compileroptions declaration true declarationdir resolve dist type verbosity usetsconfigdeclarationdir true include image postcss extensions automodules true extract index css namedexports name let reg g const temp name replace reg touppercase return temp json alias entries find src replacement resolve src noderesolve extensions some package json files have a browser field which specifies alternative files to load for people bundling for the browser if that s you either use this option or add browser to the mainfields option otherwise pkg browser will be ignored browser true preferbuiltins true mainfields replace process env node env json stringify development process env semi icon lazy load true preventassignment true autoexternal packagepath resolve importcss input resolve dist type src lib index d ts external id try const idsourcepath require resolve id paths return idsourcepath idsourcepath includes node modules catch error return false output plugins tsconfig json compileroptions target module esnext lib sourcemap true allowjs true jsx react moduleresolution node experimentaldecorators true rootdir baseurl src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true noemit true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu editor plugins transform ts optchain transform typeroots exclude package json json plugin output with verbosity text users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx → dist index esm js built in options overrides noemithelpers false importhelpers true noresolve false noemit false inlinesourcemap false outdir users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin placeholder moduleresolution allownontsextensions true parsed tsconfig options isolatedmodules false declaration true target module lib lib d ts lib dom d ts sourcemap true allowjs true jsx moduleresolution experimentaldecorators true rootdir users bytedance public helpdesk semi ui packages kefucascadertable baseurl users bytedance public helpdesk semi ui packages kefucascadertable src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu tag table src lib plugins transform ts optchain transform declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type configfilepath users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json pathsbasepath users bytedance public helpdesk semi ui packages kefucascadertable noemithelpers false noresolve false noemit false inlinesourcemap false outdir users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin placeholder allownontsextensions true filenames users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx typeacquisition enable false include exclude raw files src lib index tsx compileroptions isolatedmodules false declaration true target module esnext lib dom sourcemap true allowjs true jsx react moduleresolution node experimentaldecorators true rootdir baseurl src forceconsistentcasinginfilenames true noimplicitreturns true noimplicitthis false noimplicitany false importhelpers true strictnullchecks false suppressimplicitanyindexerrors true nounusedlocals true allowsyntheticdefaultimports true esmoduleinterop false paths ies kefu tag table src lib plugins transform ts optchain transform declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type typeroots node node modules types src typings exclude node modules compileonsave false errors wildcarddirectories compileonsave false typescript version tslib version rollup version rollup plugin version plugin options check false tsconfig users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json tsconfigoverride compileroptions declaration true declarationdir users bytedance public helpdesk semi ui packages kefucascadertable dist type verbosity usetsconfigdeclarationdir true include ts x ts x js x js x clean false cacheroot users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin exclude d ts d ts abortonerror true rollupcommonjsresolvehack false transformers tsconfigdefaults objecthashignoreunknownhack false cwd users bytedance public helpdesk semi ui packages kefucascadertable typescript version rollup config input users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx plugins name name image name postcss name json name alias name node resolve name replace name auto external name stdin output file users bytedance public helpdesk semi ui packages kefucascadertable dist index esm js format es plugins sourcemap true tsconfig path users bytedance public helpdesk semi ui packages kefucascadertable tsconfig json included ts x ts x js x js x excluded d ts d ts ambient types users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel core index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel generator index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel template index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types babel traverse index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types color name index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types css modules loader core index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types estree index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types graceful fs index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types hast index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul lib coverage index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul lib report index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types istanbul reports index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types jest index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types json schema index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types mdast index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types node index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types normalize package data index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types parse json index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types prettier index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types prop types index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types q index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types react index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types react dom index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types resolve index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types stack utils index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types unist index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types yargs index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types yargs parser index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types json schema index d ts users bytedance node modules types index d ts users bytedance node modules types minimist index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types node index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types normalize package data index d ts users bytedance public helpdesk semi ui packages kefucascadertable node modules types parse json index d ts ambient types changed redoing all semantic diagnostics transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache miss generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx dependency users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx resolving utils imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts dependency users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx resolving mock imported by users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache hit generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts transpiling users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts cache users bytedance public helpdesk semi ui packages kefucascadertable node modules cache rollup plugin code cache cache hit generated declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts generating target rolling caches emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib index tsx to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib index d ts emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib mock ts to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib mock d ts emitting declarations for users bytedance public helpdesk semi ui packages kefucascadertable src lib utils ts to users bytedance public helpdesk semi ui packages kefucascadertable dist type src lib utils d ts broken sourcemap plugins that transform code such as at position should generate accompanying sourcemaps created dist index esm js in user dist type src lib index d ts → dist index d ts error could not resolve component from dist type src lib index d ts error could not resolve component from dist type src lib index d ts | 0 |
2,862 | 28,365,904,500 | IssuesEvent | 2023-04-12 13:53:15 | cosmos/ibc-rs | https://api.github.com/repos/cosmos/ibc-rs | closed | [ICS02] `ClientStatePath` is not checked during client creation | O: reliability I: specs | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Problem Statment
- The [ICS02](https://github.com/cosmos/ibc/tree/main/spec/core/ics-002-client-semantics#create) specifies that for handling a client creation request, the process should check if the submitted `ClientState` already exists in the `ClientStatePath`!
> abortTransactionUnless(provableStore.get(clientStatePath(identifier)) === null)
- Note that an incorrect use of `SetClientState` can result in a `ClientState` being set to the same store as the one that is going to be set by a `MsgCreateClient`.
## In IBC-go
- There is a [`GetClientState`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/keeper.go#L79-L89) method, but **not** used to perform this check. Instead they create an isolated store by [`StoreClient`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/client.go#L32) for this purpose.
- This can't be applied to IBC-rs, as we do not access to the store.
## Acceptance Criteria
Check `ClientStatePath` to ensure the received `ClientState` via a `MsgCreateClient` does not exist | True | [ICS02] `ClientStatePath` is not checked during client creation - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Problem Statment
- The [ICS02](https://github.com/cosmos/ibc/tree/main/spec/core/ics-002-client-semantics#create) specifies that for handling a client creation request, the process should check if the submitted `ClientState` already exists in the `ClientStatePath`!
> abortTransactionUnless(provableStore.get(clientStatePath(identifier)) === null)
- Note that an incorrect use of `SetClientState` can result in a `ClientState` being set to the same store as the one that is going to be set by a `MsgCreateClient`.
## In IBC-go
- There is a [`GetClientState`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/keeper.go#L79-L89) method, but **not** used to perform this check. Instead they create an isolated store by [`StoreClient`](https://github.com/cosmos/ibc-go/blob/d840c699d988e32cb55f2c289d69f8ef2e5cea50/modules/core/02-client/keeper/client.go#L32) for this purpose.
- This can't be applied to IBC-rs, as we do not access to the store.
## Acceptance Criteria
Check `ClientStatePath` to ensure the received `ClientState` via a `MsgCreateClient` does not exist | reli | clientstatepath is not checked during client creation ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ problem statment the specifies that for handling a client creation request the process should check if the submitted clientstate already exists in the clientstatepath aborttransactionunless provablestore get clientstatepath identifier null note that an incorrect use of setclientstate can result in a clientstate being set to the same store as the one that is going to be set by a msgcreateclient in ibc go there is a method but not used to perform this check instead they create an isolated store by for this purpose this can t be applied to ibc rs as we do not access to the store acceptance criteria check clientstatepath to ensure the received clientstate via a msgcreateclient does not exist | 1 |
20,356 | 13,879,441,325 | IssuesEvent | 2020-10-17 14:28:20 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | closed | Error when starting local dev env | Pod: WP & Infra Status: Not Applicable Type: Bug Type: Infrastructure | ## Bug Description
When starting local dev environment I get the following error message.
```
STATUS: Import default set of media assets...
Imported file '/var/www/html/wp-content/e2e-assets/example-1.jpg' as attachment ID 1965.
Success: Imported 1 of 1 items.
Imported file '/var/www/html/wp-content/e2e-assets/example-2.jpg' as attachment ID 1966.
Success: Imported 1 of 1 items.
Imported file '/var/www/html/wp-content/e2e-assets/example-3.png' as attachment ID 1967.
Success: Imported 1 of 1 items.
STATUS: Current WordPress version: 5.6-alpha-49085...
STATUS: Updating WordPress to the latest version...
[05-Oct-2020 11:25:06 UTC] PHP Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30
Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30
```
## Expected Behaviour
It works.
## Steps to Reproduce
Type `npm run env:start`
## Screenshots
<!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! -->
## Additional Context
<!-- Please complete the following information. -->
- Plugin Version: 1-RC1-1.0.0
- Operating System: Mac OS 10.15.7
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| 1.0 | Error when starting local dev env - ## Bug Description
When starting local dev environment I get the following error message.
```
STATUS: Import default set of media assets...
Imported file '/var/www/html/wp-content/e2e-assets/example-1.jpg' as attachment ID 1965.
Success: Imported 1 of 1 items.
Imported file '/var/www/html/wp-content/e2e-assets/example-2.jpg' as attachment ID 1966.
Success: Imported 1 of 1 items.
Imported file '/var/www/html/wp-content/e2e-assets/example-3.png' as attachment ID 1967.
Success: Imported 1 of 1 items.
STATUS: Current WordPress version: 5.6-alpha-49085...
STATUS: Updating WordPress to the latest version...
[05-Oct-2020 11:25:06 UTC] PHP Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30
Warning: Declaration of WP_CLI\Core\CoreUpgrader::download_package($package, $check_signatures = true) should be compatible with WP_Upgrader::download_package($package, $check_signatures = false, $hook_extra = Array) in phar:///usr/local/bin/wp/vendor/wp-cli/core-command/src/WP_CLI/Core/CoreUpgrader.php on line 30
```
## Expected Behaviour
It works.
## Steps to Reproduce
Type `npm run env:start`
## Screenshots
<!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! -->
## Additional Context
<!-- Please complete the following information. -->
- Plugin Version: 1-RC1-1.0.0
- Operating System: Mac OS 10.15.7
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| non_reli | error when starting local dev env bug description when starting local dev environment i get the following error message status import default set of media assets imported file var www html wp content assets example jpg as attachment id success imported of items imported file var www html wp content assets example jpg as attachment id success imported of items imported file var www html wp content assets example png as attachment id success imported of items status current wordpress version alpha status updating wordpress to the latest version php warning declaration of wp cli core coreupgrader download package package check signatures true should be compatible with wp upgrader download package package check signatures false hook extra array in phar usr local bin wp vendor wp cli core command src wp cli core coreupgrader php on line warning declaration of wp cli core coreupgrader download package package check signatures true should be compatible with wp upgrader download package package check signatures false hook extra array in phar usr local bin wp vendor wp cli core command src wp cli core coreupgrader php on line expected behaviour it works steps to reproduce type npm run env start screenshots additional context plugin version operating system mac os do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief | 0 |
392,678 | 11,594,442,126 | IssuesEvent | 2020-02-24 15:20:31 | arfc/2020-dotson-optimal-sizing | https://api.github.com/repos/arfc/2020-dotson-optimal-sizing | opened | Convert steam flow rates to thermal energy requirement. | Difficulty:1-Beginner Priority:2-Normal Status:4-In Progress Type:Feature | This issue can be closed when
- [ ] mass flow rate of steam has been converted to thermal energy requirements
- [ ] sythetic histories have been generated | 1.0 | Convert steam flow rates to thermal energy requirement. - This issue can be closed when
- [ ] mass flow rate of steam has been converted to thermal energy requirements
- [ ] sythetic histories have been generated | non_reli | convert steam flow rates to thermal energy requirement this issue can be closed when mass flow rate of steam has been converted to thermal energy requirements sythetic histories have been generated | 0 |
272 | 5,938,939,826 | IssuesEvent | 2017-05-25 02:27:54 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | VS Crash: contract failed in CodeModel.AbstractProjectCodeModel.GetCodeModelCache | Area-IDE Bug Tenet-Reliability | While working in Compilers.sln.
Build:
D15Rel 26202.1
Stack trace:
```
at Roslyn.Utilities.Contract.ThrowIfFalse(Boolean condition, String message) in E:\A\_work\6\s\src\Workspaces\Core\Portable\Utilities\Contract.cs:line 86
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.GetCodeModelCache() in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 54
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.TryGetCachedFileCodeModel(String fileName, ComHandle`2& fileCodeModelHandle) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 69
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.FireEvents(DocumentId documentId, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 86
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.AnalyzeSyntaxAsync(Document document, InvocationReasons reasons, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 60
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass30_0.<ProcessDocumentAnalyzersAsync>b__0(IIncrementalAnalyzer a, Document d, CancellationToken c) in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 153
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass31_1`1.<<RunAnalyzersAsync>b__0>d.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 187
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<GetOrDefaultAsync>d__33`2.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 229
```
Memory dump:
```%internal_share%\public\tomat\Bugs\16973```
| True | VS Crash: contract failed in CodeModel.AbstractProjectCodeModel.GetCodeModelCache - While working in Compilers.sln.
Build:
D15Rel 26202.1
Stack trace:
```
at Roslyn.Utilities.Contract.ThrowIfFalse(Boolean condition, String message) in E:\A\_work\6\s\src\Workspaces\Core\Portable\Utilities\Contract.cs:line 86
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.GetCodeModelCache() in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 54
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.AbstractProjectCodeModel.TryGetCachedFileCodeModel(String fileName, ComHandle`2& fileCodeModelHandle) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\AbstractProjectCodeModel.cs:line 69
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.FireEvents(DocumentId documentId, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 86
at Microsoft.VisualStudio.LanguageServices.Implementation.CodeModel.CodeModelIncrementalAnalyzerProvider.Analyzer.AnalyzeSyntaxAsync(Document document, InvocationReasons reasons, CancellationToken cancellationToken) in E:\A\_work\6\s\src\VisualStudio\Core\Impl\CodeModel\CodeModelIncrementalAnalyzer.cs:line 60
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass30_0.<ProcessDocumentAnalyzersAsync>b__0(IIncrementalAnalyzer a, Document d, CancellationToken c) in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 153
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<>c__DisplayClass31_1`1.<<RunAnalyzersAsync>b__0>d.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 187
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.CodeAnalysis.SolutionCrawler.SolutionCrawlerRegistrationService.WorkCoordinator.IncrementalAnalyzerProcessor.<GetOrDefaultAsync>d__33`2.MoveNext() in E:\A\_work\6\s\src\Features\Core\Portable\SolutionCrawler\WorkCoordinator.IncrementalAnalyzerProcessor.cs:line 229
```
Memory dump:
```%internal_share%\public\tomat\Bugs\16973```
| reli | vs crash contract failed in codemodel abstractprojectcodemodel getcodemodelcache while working in compilers sln build stack trace at roslyn utilities contract throwiffalse boolean condition string message in e a work s src workspaces core portable utilities contract cs line at microsoft visualstudio languageservices implementation codemodel abstractprojectcodemodel getcodemodelcache in e a work s src visualstudio core impl codemodel abstractprojectcodemodel cs line at microsoft visualstudio languageservices implementation codemodel abstractprojectcodemodel trygetcachedfilecodemodel string filename comhandle filecodemodelhandle in e a work s src visualstudio core impl codemodel abstractprojectcodemodel cs line at microsoft visualstudio languageservices implementation codemodel codemodelincrementalanalyzerprovider analyzer fireevents documentid documentid cancellationtoken cancellationtoken in e a work s src visualstudio core impl codemodel codemodelincrementalanalyzer cs line at microsoft visualstudio languageservices implementation codemodel codemodelincrementalanalyzerprovider analyzer analyzesyntaxasync document document invocationreasons reasons cancellationtoken cancellationtoken in e a work s src visualstudio core impl codemodel codemodelincrementalanalyzer cs line at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor c b iincrementalanalyzer a document d cancellationtoken c in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor c b d movenext in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft codeanalysis solutioncrawler solutioncrawlerregistrationservice workcoordinator incrementalanalyzerprocessor d movenext in e a work s src features core portable solutioncrawler workcoordinator incrementalanalyzerprocessor cs line memory dump internal share public tomat bugs | 1 |
1,374 | 15,606,232,799 | IssuesEvent | 2021-03-19 07:42:05 | FoundationDB/fdb-kubernetes-operator | https://api.github.com/repos/FoundationDB/fdb-kubernetes-operator | closed | Logging all decisions to remove processes | reliability | We've seen some cases where processes were marked for removal and it wasn't clear why. We should add a log statement whenever we mark a process for removal identifying the reason, so we can search for this event in the logs when we see unexplained removals. | True | Logging all decisions to remove processes - We've seen some cases where processes were marked for removal and it wasn't clear why. We should add a log statement whenever we mark a process for removal identifying the reason, so we can search for this event in the logs when we see unexplained removals. | reli | logging all decisions to remove processes we ve seen some cases where processes were marked for removal and it wasn t clear why we should add a log statement whenever we mark a process for removal identifying the reason so we can search for this event in the logs when we see unexplained removals | 1 |
2,743 | 27,378,169,223 | IssuesEvent | 2023-02-28 08:05:41 | jasp-stats/jasp-issues | https://api.github.com/repos/jasp-stats/jasp-issues | closed | Intercalas correlation coefficient | Module: jaspReliability |
* Enhancement: add the ICC calculation in the reliability analysis section
* Purpose: to calculate the ICC for reliability studies with multiple quantitative variables
* Use-case: in sports science is typically asses the reliability of instruments that report quantitative variables, nowadays we need to compute this calculation using other statistical packages like SPSS.
| True | Intercalas correlation coefficient -
* Enhancement: add the ICC calculation in the reliability analysis section
* Purpose: to calculate the ICC for reliability studies with multiple quantitative variables
* Use-case: in sports science is typically asses the reliability of instruments that report quantitative variables, nowadays we need to compute this calculation using other statistical packages like SPSS.
| reli | intercalas correlation coefficient enhancement add the icc calculation in the reliability analysis section purpose to calculate the icc for reliability studies with multiple quantitative variables use case in sports science is typically asses the reliability of instruments that report quantitative variables nowadays we need to compute this calculation using other statistical packages like spss | 1 |
1,581 | 17,263,685,794 | IssuesEvent | 2021-07-22 11:07:24 | ppy/osu | https://api.github.com/repos/ppy/osu | closed | Game crashes when entering playlists lounge the second time on dev server | type:reliability | **Describe the bug:**
1. Enter the playlists lounge
2. Quit to main menu
3. Enter the playlists lounge again
4. Game crashes when loading the lounge screen for the second time.
https://user-images.githubusercontent.com/25472513/126591824-e10a3958-39b9-4c8e-8c5e-3c8dc71fe648.mp4
The dev account I'm using is completely new (just created today, haven't played any maps on it).
**osu!lazer version:** master (edce3e0efeaeff1d8479374e834280b2b8bc4c2f)
**Logs:**
[performance.log](https://github.com/ppy/osu/files/6860086/performance.log)
[performance-audio.log](https://github.com/ppy/osu/files/6860087/performance-audio.log)
[performance-draw.log](https://github.com/ppy/osu/files/6860088/performance-draw.log)
[performance-input.log](https://github.com/ppy/osu/files/6860080/performance-input.log)
[performance-update.log](https://github.com/ppy/osu/files/6860081/performance-update.log)
[runtime.log](https://github.com/ppy/osu/files/6860082/runtime.log)
[updater.log](https://github.com/ppy/osu/files/6860083/updater.log)
[database.log](https://github.com/ppy/osu/files/6860084/database.log)
[network.log](https://github.com/ppy/osu/files/6860085/network.log)
| True | Game crashes when entering playlists lounge the second time on dev server - **Describe the bug:**
1. Enter the playlists lounge
2. Quit to main menu
3. Enter the playlists lounge again
4. Game crashes when loading the lounge screen for the second time.
https://user-images.githubusercontent.com/25472513/126591824-e10a3958-39b9-4c8e-8c5e-3c8dc71fe648.mp4
The dev account I'm using is completely new (just created today, haven't played any maps on it).
**osu!lazer version:** master (edce3e0efeaeff1d8479374e834280b2b8bc4c2f)
**Logs:**
[performance.log](https://github.com/ppy/osu/files/6860086/performance.log)
[performance-audio.log](https://github.com/ppy/osu/files/6860087/performance-audio.log)
[performance-draw.log](https://github.com/ppy/osu/files/6860088/performance-draw.log)
[performance-input.log](https://github.com/ppy/osu/files/6860080/performance-input.log)
[performance-update.log](https://github.com/ppy/osu/files/6860081/performance-update.log)
[runtime.log](https://github.com/ppy/osu/files/6860082/runtime.log)
[updater.log](https://github.com/ppy/osu/files/6860083/updater.log)
[database.log](https://github.com/ppy/osu/files/6860084/database.log)
[network.log](https://github.com/ppy/osu/files/6860085/network.log)
| reli | game crashes when entering playlists lounge the second time on dev server describe the bug enter the playlists lounge quit to main menu enter the playlists lounge again game crashes when loading the lounge screen for the second time the dev account i m using is completely new just created today haven t played any maps on it osu lazer version master logs | 1 |
930 | 11,707,171,881 | IssuesEvent | 2020-03-08 04:03:17 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Dependency on System.Runtime.InteropServices.RuntimeInformation causes runtime error for apps targeting net451 on *nix systems (i.e. running on Mono) | area-System.Runtime enhancement tenet-reliability untriaged | ## Scenario
1) Running a `net451` app depending on `System.Runtime.InteropServices.RuntimeInformation` on OSX/Ubuntu will produce the following exception at runtime:
```
dotnet -v run -f net451 1 ↵
Telemetry is: Enabled
Project RuntimeServiceOnMono (.NETFramework,Version=v4.5.1) was previously compiled. Skipping compilation.
Running /usr/local/bin/mono --debug /Users/jtluo/Documents/workspace/juntaoluo/tp/RuntimeServiceOnMono/bin/Debug/net451/osx.10.11-x64/RuntimeServiceOnMono.exe
Process ID: 29130
Unhandled Exception:
System.DllNotFoundException: System.Native
at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate ()
at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0
at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0
at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.DllNotFoundException: System.Native
at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate ()
at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0
at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0
at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0
```
## Example
See repro at https://github.com/JunTaoLuo/RuntimeServiceOnMono. To run the sample, run `dotnet restore` and `dotnet run -f net451`
I understand Mono is not a scenario that's actively being developed for but we should not be causing a exception like this especially since we intend to use the `InteropServices` APIs in often used ASP.NET packages like Logging.
Currently this means that most of our samples and apps cannot run on Mono (full CLR on *nix).
## Environment
`dotnet --version`:
```
Microsoft .NET Core Shared Framework Host
Version : 1.0.1-rc3-004312-00
Build : 1db6c07638a70a621b312e78d4dc9fb31a530f2f
```
`mono -V`:
```
Mono JIT compiler version 4.0.5 ((detached/1d8d582 Tue Oct 20 15:15:33 EDT 2015)
Copyright (C) 2002-2014 Novell, Inc, Xamarin Inc and Contributors. www.mono-project.com
TLS: normal
SIGSEGV: altstack
Notification: kqueue
Architecture: x86
Disabled: none
Misc: softdebug
LLVM: yes(3.6.0svn-mono-(detached/a173357)
GC: sgen
```
cc @BrennanConroy @muratg @Eilon
| True | Dependency on System.Runtime.InteropServices.RuntimeInformation causes runtime error for apps targeting net451 on *nix systems (i.e. running on Mono) - ## Scenario
1) Running a `net451` app depending on `System.Runtime.InteropServices.RuntimeInformation` on OSX/Ubuntu will produce the following exception at runtime:
```
dotnet -v run -f net451 1 ↵
Telemetry is: Enabled
Project RuntimeServiceOnMono (.NETFramework,Version=v4.5.1) was previously compiled. Skipping compilation.
Running /usr/local/bin/mono --debug /Users/jtluo/Documents/workspace/juntaoluo/tp/RuntimeServiceOnMono/bin/Debug/net451/osx.10.11-x64/RuntimeServiceOnMono.exe
Process ID: 29130
Unhandled Exception:
System.DllNotFoundException: System.Native
at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate ()
at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0
at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0
at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.DllNotFoundException: System.Native
at (wrapper managed-to-native) Interop/Sys:GetUnixNamePrivate ()
at Interop+Sys.GetUnixName () [0x00000] in <filename unknown>:0
at System.Runtime.InteropServices.RuntimeInformation.IsOSPlatform (OSPlatform osPlatform) [0x00000] in <filename unknown>:0
at ConsoleApplication.Program.Main (System.String[] args) [0x00000] in <filename unknown>:0
```
## Example
See repro at https://github.com/JunTaoLuo/RuntimeServiceOnMono. To run the sample, run `dotnet restore` and `dotnet run -f net451`
I understand Mono is not a scenario that's actively being developed for but we should not be causing a exception like this especially since we intend to use the `InteropServices` APIs in often used ASP.NET packages like Logging.
Currently this means that most of our samples and apps cannot run on Mono (full CLR on *nix).
## Environment
`dotnet --version`:
```
Microsoft .NET Core Shared Framework Host
Version : 1.0.1-rc3-004312-00
Build : 1db6c07638a70a621b312e78d4dc9fb31a530f2f
```
`mono -V`:
```
Mono JIT compiler version 4.0.5 ((detached/1d8d582 Tue Oct 20 15:15:33 EDT 2015)
Copyright (C) 2002-2014 Novell, Inc, Xamarin Inc and Contributors. www.mono-project.com
TLS: normal
SIGSEGV: altstack
Notification: kqueue
Architecture: x86
Disabled: none
Misc: softdebug
LLVM: yes(3.6.0svn-mono-(detached/a173357)
GC: sgen
```
cc @BrennanConroy @muratg @Eilon
| reli | dependency on system runtime interopservices runtimeinformation causes runtime error for apps targeting on nix systems i e running on mono scenario running a app depending on system runtime interopservices runtimeinformation on osx ubuntu will produce the following exception at runtime dotnet v run f ↵ telemetry is enabled project runtimeserviceonmono netframework version was previously compiled skipping compilation running usr local bin mono debug users jtluo documents workspace juntaoluo tp runtimeserviceonmono bin debug osx runtimeserviceonmono exe process id unhandled exception system dllnotfoundexception system native at wrapper managed to native interop sys getunixnameprivate at interop sys getunixname in at system runtime interopservices runtimeinformation isosplatform osplatform osplatform in at consoleapplication program main system string args in fatal unhandled exception system dllnotfoundexception system native at wrapper managed to native interop sys getunixnameprivate at interop sys getunixname in at system runtime interopservices runtimeinformation isosplatform osplatform osplatform in at consoleapplication program main system string args in example see repro at to run the sample run dotnet restore and dotnet run f i understand mono is not a scenario that s actively being developed for but we should not be causing a exception like this especially since we intend to use the interopservices apis in often used asp net packages like logging currently this means that most of our samples and apps cannot run on mono full clr on nix environment dotnet version microsoft net core shared framework host version build mono v mono jit compiler version detached tue oct edt copyright c novell inc xamarin inc and contributors tls normal sigsegv altstack notification kqueue architecture disabled none misc softdebug llvm yes mono detached gc sgen cc brennanconroy muratg eilon | 1 |
450 | 7,612,728,123 | IssuesEvent | 2018-05-01 18:35:57 | m3db/m3db | https://api.github.com/repos/m3db/m3db | opened | Monitor file sizes | C: Toolchain P: Medium T: Instrumentation T: Reliability | Add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard/index level.
i.e.
```
/data/namespace/shard <-- data fileset files
/snapshot/namespace/shard <-- snapshot fileset files
```
| True | Monitor file sizes - Add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard/index level.
i.e.
```
/data/namespace/shard <-- data fileset files
/snapshot/namespace/shard <-- snapshot fileset files
```
| reli | monitor file sizes add a bg routine which periodically emits metrics about size of directories for the the various fileset files at the shard index level i e data namespace shard data fileset files snapshot namespace shard snapshot fileset files | 1 |
271,689 | 8,488,531,242 | IssuesEvent | 2018-10-26 16:56:07 | FIDUCEO/FCDR_HIRS | https://api.github.com/repos/FIDUCEO/FCDR_HIRS | closed | Odd values for channel 4 and 12 on metopa | Priority: High |

particularly clear in time series

should investigate!
Compare also Shi and Bates (2011). | 1.0 | Odd values for channel 4 and 12 on metopa -

particularly clear in time series

should investigate!
Compare also Shi and Bates (2011). | non_reli | odd values for channel and on metopa particularly clear in time series should investigate compare also shi and bates | 0 |
35,914 | 2,793,820,595 | IssuesEvent | 2015-05-11 13:37:31 | elecoest/allevents-3-2 | https://api.github.com/repos/elecoest/allevents-3-2 | closed | FrontEnd - Amélioration - Rajout page affichage des lieux et événements géolocalisés | auto-migrated Priority-Medium Type-Enhancement | ```
Dans AE on dispose d'une liste de lieux qu'il serait intéressant de pouvoir
visualiser en front sous forme d'une carte.
En effet, les lieux sont un peu ce qui différencie AE de certaines autres
extensions de gestion d'agendas qui n'en possèdent pas.
Pouvoir les mettre en valeur (et les utiliser) en front me semble couler de
source.
Il existe bien des composants spécialisés dans l'affichage sous forme de
carte de lieux ou de contacts mais si on voulait les utiliser avec AE, il
faudrait alors re-saisir les lieux (latitude, longitude, titres ...) dans ces
composants ce qui me semble un peu dommage.
Dans AE la page d'affichage de carte pourrait avoir plusieurs modes
d'affichages :
- Carte des Lieux
- Carte des événements (par lieu et par date).
Avec filtres sur agendas, catégories etc... en fonction du choix et des
possibilités de filtrage des éléments choisis.
Sur la carte à la place d'une icône par défaut, les vignettes pourraient
être utilisées.
Au survol de chaque éléments, une fenêtre de type Pop-up pourrait afficher
le détail de l'élément survolé. Si une adresse mail est disponible, un clic
sur le titre de l'élément (dans la fenêtre pop-up) permettrait d'écrire au
responsable du lieux ou de l'événement (choix de permettre ou non cette
fonction dans les paramètres de la page).
Un exemple ci-dessous.
```
Original issue reported on code.google.com by `jjacquesh` on 13 Dec 2014 at 11:20
Attachments:
* [20141213_cartographie_des_lieux.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-302/comment-0/20141213_cartographie_des_lieux.jpg)
| 1.0 | FrontEnd - Amélioration - Rajout page affichage des lieux et événements géolocalisés - ```
Dans AE on dispose d'une liste de lieux qu'il serait intéressant de pouvoir
visualiser en front sous forme d'une carte.
En effet, les lieux sont un peu ce qui différencie AE de certaines autres
extensions de gestion d'agendas qui n'en possèdent pas.
Pouvoir les mettre en valeur (et les utiliser) en front me semble couler de
source.
Il existe bien des composants spécialisés dans l'affichage sous forme de
carte de lieux ou de contacts mais si on voulait les utiliser avec AE, il
faudrait alors re-saisir les lieux (latitude, longitude, titres ...) dans ces
composants ce qui me semble un peu dommage.
Dans AE la page d'affichage de carte pourrait avoir plusieurs modes
d'affichages :
- Carte des Lieux
- Carte des événements (par lieu et par date).
Avec filtres sur agendas, catégories etc... en fonction du choix et des
possibilités de filtrage des éléments choisis.
Sur la carte à la place d'une icône par défaut, les vignettes pourraient
être utilisées.
Au survol de chaque éléments, une fenêtre de type Pop-up pourrait afficher
le détail de l'élément survolé. Si une adresse mail est disponible, un clic
sur le titre de l'élément (dans la fenêtre pop-up) permettrait d'écrire au
responsable du lieux ou de l'événement (choix de permettre ou non cette
fonction dans les paramètres de la page).
Un exemple ci-dessous.
```
Original issue reported on code.google.com by `jjacquesh` on 13 Dec 2014 at 11:20
Attachments:
* [20141213_cartographie_des_lieux.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-302/comment-0/20141213_cartographie_des_lieux.jpg)
| non_reli | frontend amélioration rajout page affichage des lieux et événements géolocalisés dans ae on dispose d une liste de lieux qu il serait intéressant de pouvoir visualiser en front sous forme d une carte en effet les lieux sont un peu ce qui différencie ae de certaines autres extensions de gestion d agendas qui n en possèdent pas pouvoir les mettre en valeur et les utiliser en front me semble couler de source il existe bien des composants spécialisés dans l affichage sous forme de carte de lieux ou de contacts mais si on voulait les utiliser avec ae il faudrait alors re saisir les lieux latitude longitude titres dans ces composants ce qui me semble un peu dommage dans ae la page d affichage de carte pourrait avoir plusieurs modes d affichages carte des lieux carte des événements par lieu et par date avec filtres sur agendas catégories etc en fonction du choix et des possibilités de filtrage des éléments choisis sur la carte à la place d une icône par défaut les vignettes pourraient être utilisées au survol de chaque éléments une fenêtre de type pop up pourrait afficher le détail de l élément survolé si une adresse mail est disponible un clic sur le titre de l élément dans la fenêtre pop up permettrait d écrire au responsable du lieux ou de l événement choix de permettre ou non cette fonction dans les paramètres de la page un exemple ci dessous original issue reported on code google com by jjacquesh on dec at attachments | 0 |
2,489 | 25,770,265,519 | IssuesEvent | 2022-12-09 07:15:21 | juliuspf/jasp-issues | https://api.github.com/repos/juliuspf/jasp-issues | closed | oh new please | Module: jaspReliability Feature Request | ### Description
_No response_
### Purpose
_No response_
### Use-case
_No response_
### Is your feature request related to a problem?
_No response_
### Is your feature request related to a JASP module?
Reliability
### Describe the solution you would like
_No response_
### Describe alternatives that you have considered
_No response_
### Additional context
_No response_ | True | oh new please - ### Description
_No response_
### Purpose
_No response_
### Use-case
_No response_
### Is your feature request related to a problem?
_No response_
### Is your feature request related to a JASP module?
Reliability
### Describe the solution you would like
_No response_
### Describe alternatives that you have considered
_No response_
### Additional context
_No response_ | reli | oh new please description no response purpose no response use case no response is your feature request related to a problem no response is your feature request related to a jasp module reliability describe the solution you would like no response describe alternatives that you have considered no response additional context no response | 1 |
1,682 | 18,518,547,317 | IssuesEvent | 2021-10-20 12:55:00 | hashicorp/consul | https://api.github.com/repos/hashicorp/consul | closed | docs: Note the Grafana dashboard | type/docs good first issue theme/telemetry theme/reliability Hacktoberfest | Since last year we've had an "official" Grafana dashboard which lives in the source tree and we [publish publicly](https://grafana.com/grafana/dashboards/13396).
But many users including internal teams aren't aware it exists.
We should update the Telemetry page to note it's existence! | True | docs: Note the Grafana dashboard - Since last year we've had an "official" Grafana dashboard which lives in the source tree and we [publish publicly](https://grafana.com/grafana/dashboards/13396).
But many users including internal teams aren't aware it exists.
We should update the Telemetry page to note it's existence! | reli | docs note the grafana dashboard since last year we ve had an official grafana dashboard which lives in the source tree and we but many users including internal teams aren t aware it exists we should update the telemetry page to note it s existence | 1 |
1,532 | 16,759,295,826 | IssuesEvent | 2021-06-13 13:01:35 | skytable/skytable | https://api.github.com/repos/skytable/skytable | closed | Server: Running two skyd in same directory can lead to loss of data | A-independent C-bug C-reliability C-storage D-server P-high S-waiting-on-author | **Description of bug**
Currently it is possible to start two `skyd` processes from the same directory. This will lead to data being overwritten by the other process.
**Steps to reproduce**
Steps to reproduce the behavior:
For example:
1. Run `skyd`
2. Run `skyd` again
3. Store some data on both servers
3. When both `skyd`are closed you will end up with data for only the last server closed
**Expected behavior**
Stop any subsequent skyd process from starting up if there is already one running in that directory.
**Meta**
- **Release tag:**
None
- **Branch:**
- **Commit ID:**
- **Operating system:**
| True | Server: Running two skyd in same directory can lead to loss of data - **Description of bug**
Currently it is possible to start two `skyd` processes from the same directory. This will lead to data being overwritten by the other process.
**Steps to reproduce**
Steps to reproduce the behavior:
For example:
1. Run `skyd`
2. Run `skyd` again
3. Store some data on both servers
3. When both `skyd`are closed you will end up with data for only the last server closed
**Expected behavior**
Stop any subsequent skyd process from starting up if there is already one running in that directory.
**Meta**
- **Release tag:**
None
- **Branch:**
- **Commit ID:**
- **Operating system:**
| reli | server running two skyd in same directory can lead to loss of data description of bug currently it is possible to start two skyd processes from the same directory this will lead to data being overwritten by the other process steps to reproduce steps to reproduce the behavior for example run skyd run skyd again store some data on both servers when both skyd are closed you will end up with data for only the last server closed expected behavior stop any subsequent skyd process from starting up if there is already one running in that directory meta release tag none branch commit id operating system | 1 |
1,799 | 19,915,685,483 | IssuesEvent | 2022-01-25 22:19:05 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | DownloadToFile causes an error - reactor.core.Exceptions$OverflowException & java.util.concurrent.TimeoutException | question Storage Client customer-reported pillar-reliability | 2021-10-21 16:34:04,381 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:04,383 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:181) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:04,410 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:20,566 ERROR [Thread-55] (AzureDownloadToLocalFileCopy:62) - Azure copyFile failed : java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
reactor.core.Exceptions$ReactiveException: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
at reactor.core.Exceptions.propagate(Exceptions.java:392) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:97) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?]
at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?]
at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?]
at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?]
at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?]
at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?]
Suppressed: java.lang.Exception: #block terminated with an error
at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?]
at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?]
at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?]
at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?]
at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?]
at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?]
Caused by: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.handleTimeout(FluxTimeout.java:294) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.doTimeout(FluxTimeout.java:279) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutTimeoutSubscriber.onNext(FluxTimeout.java:418) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.MonoDelay$MonoDelayRunnable.propagateDelay(MonoDelay.java:270) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.MonoDelay$MonoDelayRunnable.run(MonoDelay.java:285) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28) ~[reactor-core-3.4.6.jar!/:3.4.6]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?] | True | DownloadToFile causes an error - reactor.core.Exceptions$OverflowException & java.util.concurrent.TimeoutException - 2021-10-21 16:34:04,381 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:04,383 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:181) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:04,410 DEBUG [Thread-64] (Operators:254) - More data produced than requested
reactor.core.Exceptions$OverflowException: The receiver is overrun by more signals than expected (bounded queue...)
at reactor.core.Exceptions.failWithOverflow(Exceptions.java:220) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators.reportMoreProduced(Operators.java:1071) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.producedOne(Operators.java:2112) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:82) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:120) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.onNext(FluxTimeout.java:179) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:130) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:353) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:118) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:171) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil.lambda$toFluxByteBuffer$2(FluxUtil.java:227) ~[azure-core-1.17.0.jar!/:?]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:264) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:206) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.request(FluxFilterFuseable.java:191) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.request(FluxUsing.java:320) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.request(FluxDoFinally.java:157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.SerializedSubscriber.request(SerializedSubscriber.java:151) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxMap$MapSubscriber.request(FluxMap.java:162) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.request(Operators.java:2157) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxPeek$PeekSubscriber.request(FluxPeek.java:137) [reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.StrictSubscriber.request(StrictSubscriber.java:138) [reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:444) [azure-core-1.17.0.jar!/:?]
at com.azure.core.util.FluxUtil$1$1.completed(FluxUtil.java:435) [azure-core-1.17.0.jar!/:?]
at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:127) [?:?]
at sun.nio.ch.SimpleAsynchronousFileChannelImpl$3.run(SimpleAsynchronousFileChannelImpl.java:389) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
2021-10-21 16:34:20,566 ERROR [Thread-55] (AzureDownloadToLocalFileCopy:62) - Azure copyFile failed : java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
reactor.core.Exceptions$ReactiveException: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
at reactor.core.Exceptions.propagate(Exceptions.java:392) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:97) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?]
at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?]
at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?]
at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?]
at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?]
at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?]
Suppressed: java.lang.Exception: #block terminated with an error
at reactor.core.publisher.BlockingSingleSubscriber.blockingGet(BlockingSingleSubscriber.java:99) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.Mono.block(Mono.java:1703) ~[reactor-core-3.4.6.jar!/:3.4.6]
at com.azure.storage.common.implementation.StorageImplUtils.blockWithOptionalTimeout(StorageImplUtils.java:128) ~[azure-storage-common-12.12.0.jar!/:?]
at com.azure.storage.blob.specialized.BlobClientBase.downloadToFileWithResponse(BlobClientBase.java:903) ~[azure-storage-blob-12.12.0.jar!/:?]
at com.bmc.cm.aft.client.wrapper.AzureWrapper.downloadBlob(AzureWrapper.java:543) ~[classes!/:?]
at com.bmc.cm.aft.client.AzureDownloadToLocalFileCopy.copyFile(AzureDownloadToLocalFileCopy.java:58) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.transfer(FileCopy.java:1601) [classes!/:?]
at com.bmc.cm.aft.client.FileCopy.run(FileCopy.java:1421) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copySingleFile(FTCTransfer.java:875) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.copy(FTCTransfer.java:770) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.runCopy(FTCTransfer.java:372) [classes!/:?]
at com.bmc.cm.aft.client.FTCTransfer.run(FTCTransfer.java:301) [classes!/:?]
at com.bmc.cm.aft.client.SubtaskRunner.run(SubtaskRunner.java:599) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runSubTasks(FTCHandler.java:400) [classes!/:?]
at com.bmc.cm.aft.client.FTCHandler.runJob(FTCHandler.java:197) [classes!/:?]
at com.bmc.cm.aft.jobmanagement.JobObjectImpl$JobRunnerThread.run(JobObjectImpl.java:366) [classes!/:?]
Caused by: java.util.concurrent.TimeoutException: Did not observe any item or terminal signal within 60000ms in 'doFinally' (and no fallback has been configured)
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.handleTimeout(FluxTimeout.java:294) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutMainSubscriber.doTimeout(FluxTimeout.java:279) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxTimeout$TimeoutTimeoutSubscriber.onNext(FluxTimeout.java:418) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.MonoDelay$MonoDelayRunnable.propagateDelay(MonoDelay.java:270) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.publisher.MonoDelay$MonoDelayRunnable.run(MonoDelay.java:285) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:68) ~[reactor-core-3.4.6.jar!/:3.4.6]
at reactor.core.scheduler.SchedulerTask.call(SchedulerTask.java:28) ~[reactor-core-3.4.6.jar!/:3.4.6]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[?:?]
at java.lang.Thread.run(Thread.java:834) ~[?:?] | reli | downloadtofile causes an error reactor core exceptions overflowexception java util concurrent timeoutexception debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmap mapsubscriber onnext fluxmap java at reactor core publisher serializedsubscriber onnext serializedsubscriber java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java debug operators more data produced than requested reactor core exceptions overflowexception the receiver is overrun by more signals than expected bounded queue at reactor core exceptions failwithoverflow exceptions java at reactor core publisher operators reportmoreproduced operators java at reactor core publisher operators multisubscriptionsubscriber producedone operators java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmap mapsubscriber onnext fluxmap java at reactor core publisher serializedsubscriber onnext serializedsubscriber java at reactor core publisher fluxtimeout timeoutmainsubscriber onnext fluxtimeout java at reactor core publisher fluxdofinally dofinallysubscriber onnext fluxdofinally java at reactor core publisher fluxusing usingfuseablesubscriber onnext fluxusing java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber onnext fluxfilterfuseable java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure core util fluxutil lambda tofluxbytebuffer fluxutil java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxfilterfuseable filterfuseablesubscriber request fluxfilterfuseable java at reactor core publisher fluxusing usingfuseablesubscriber request fluxusing java at reactor core publisher fluxdofinally dofinallysubscriber request fluxdofinally java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher serializedsubscriber request serializedsubscriber java at reactor core publisher fluxmap mapsubscriber request fluxmap java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher operators multisubscriptionsubscriber request operators java at reactor core publisher fluxpeek peeksubscriber request fluxpeek java at reactor core publisher strictsubscriber request strictsubscriber java at com azure core util fluxutil completed fluxutil java at com azure core util fluxutil completed fluxutil java at sun nio ch invoker invokeunchecked invoker java at sun nio ch simpleasynchronousfilechannelimpl run simpleasynchronousfilechannelimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java error azuredownloadtolocalfilecopy azure copyfile failed java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured reactor core exceptions reactiveexception java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured at reactor core exceptions propagate exceptions java at reactor core publisher blockingsinglesubscriber blockingget blockingsinglesubscriber java at reactor core publisher mono block mono java at com azure storage common implementation storageimplutils blockwithoptionaltimeout storageimplutils java at com azure storage blob specialized blobclientbase downloadtofilewithresponse blobclientbase java at com bmc cm aft client wrapper azurewrapper downloadblob azurewrapper java at com bmc cm aft client azuredownloadtolocalfilecopy copyfile azuredownloadtolocalfilecopy java at com bmc cm aft client filecopy transfer filecopy java at com bmc cm aft client filecopy run filecopy java at com bmc cm aft client ftctransfer copysinglefile ftctransfer java at com bmc cm aft client ftctransfer copy ftctransfer java at com bmc cm aft client ftctransfer runcopy ftctransfer java at com bmc cm aft client ftctransfer run ftctransfer java at com bmc cm aft client subtaskrunner run subtaskrunner java at com bmc cm aft client ftchandler runsubtasks ftchandler java at com bmc cm aft client ftchandler runjob ftchandler java at com bmc cm aft jobmanagement jobobjectimpl jobrunnerthread run jobobjectimpl java suppressed java lang exception block terminated with an error at reactor core publisher blockingsinglesubscriber blockingget blockingsinglesubscriber java at reactor core publisher mono block mono java at com azure storage common implementation storageimplutils blockwithoptionaltimeout storageimplutils java at com azure storage blob specialized blobclientbase downloadtofilewithresponse blobclientbase java at com bmc cm aft client wrapper azurewrapper downloadblob azurewrapper java at com bmc cm aft client azuredownloadtolocalfilecopy copyfile azuredownloadtolocalfilecopy java at com bmc cm aft client filecopy transfer filecopy java at com bmc cm aft client filecopy run filecopy java at com bmc cm aft client ftctransfer copysinglefile ftctransfer java at com bmc cm aft client ftctransfer copy ftctransfer java at com bmc cm aft client ftctransfer runcopy ftctransfer java at com bmc cm aft client ftctransfer run ftctransfer java at com bmc cm aft client subtaskrunner run subtaskrunner java at com bmc cm aft client ftchandler runsubtasks ftchandler java at com bmc cm aft client ftchandler runjob ftchandler java at com bmc cm aft jobmanagement jobobjectimpl jobrunnerthread run jobobjectimpl java caused by java util concurrent timeoutexception did not observe any item or terminal signal within in dofinally and no fallback has been configured at reactor core publisher fluxtimeout timeoutmainsubscriber handletimeout fluxtimeout java at reactor core publisher fluxtimeout timeoutmainsubscriber dotimeout fluxtimeout java at reactor core publisher fluxtimeout timeouttimeoutsubscriber onnext fluxtimeout java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher monodelay monodelayrunnable propagatedelay monodelay java at reactor core publisher monodelay monodelayrunnable run monodelay java at reactor core scheduler schedulertask call schedulertask java at reactor core scheduler schedulertask call schedulertask java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java | 1 |
1,970 | 22,293,542,308 | IssuesEvent | 2022-06-12 18:13:53 | jina-ai/jina | https://api.github.com/repos/jina-ai/jina | closed | Server side throttling | epic/reliability | At server side throttling so that Gateway/Head/Worker can reject new requests if there are too many outstanding requests already queued up. This needs to configurable | True | Server side throttling - At server side throttling so that Gateway/Head/Worker can reject new requests if there are too many outstanding requests already queued up. This needs to configurable | reli | server side throttling at server side throttling so that gateway head worker can reject new requests if there are too many outstanding requests already queued up this needs to configurable | 1 |
2,390 | 25,100,822,486 | IssuesEvent | 2022-11-08 13:31:00 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | opened | Disable the message TTL by default | kind/toil scope/clients-java scope/clients-go severity/mid area/reliability | **Description**
I can publish a message using Zeebe's Java or Go client. The message can have a time-to-live (TTL). It can be correlated within the given TTL. After the TTL is reached, the message expires and is removed.
Currently, the Java and the Go client use a different default value for the message TTL.
* in the Java client: the default TTL is [1 hour](https://github.com/camunda/zeebe/blob/main/clients/java/src/main/java/io/camunda/zeebe/client/impl/ZeebeClientBuilderImpl.java#L57)
* in the Go client: the default TTL is [5 seconds](https://github.com/camunda/zeebe/blob/main/clients/go/cmd/zbctl/internal/commands/publishMessage.go#L74)
We should align the default value in all (official) clients to align the general behavior.
I recommend setting the default message TTL to `zero` (i.e. disable the message buffering).
If the TTL is set to a high value then it can impact the performance of the broker (see https://github.com/camunda/zeebe/issues/10643#issuecomment-1303710858). Instead of setting an arbitrary value for the TTL, let the user explicitly choose to use the buffering and think about a reasonable value. | True | Disable the message TTL by default - **Description**
I can publish a message using Zeebe's Java or Go client. The message can have a time-to-live (TTL). It can be correlated within the given TTL. After the TTL is reached, the message expires and is removed.
Currently, the Java and the Go client use a different default value for the message TTL.
* in the Java client: the default TTL is [1 hour](https://github.com/camunda/zeebe/blob/main/clients/java/src/main/java/io/camunda/zeebe/client/impl/ZeebeClientBuilderImpl.java#L57)
* in the Go client: the default TTL is [5 seconds](https://github.com/camunda/zeebe/blob/main/clients/go/cmd/zbctl/internal/commands/publishMessage.go#L74)
We should align the default value in all (official) clients to align the general behavior.
I recommend setting the default message TTL to `zero` (i.e. disable the message buffering).
If the TTL is set to a high value then it can impact the performance of the broker (see https://github.com/camunda/zeebe/issues/10643#issuecomment-1303710858). Instead of setting an arbitrary value for the TTL, let the user explicitly choose to use the buffering and think about a reasonable value. | reli | disable the message ttl by default description i can publish a message using zeebe s java or go client the message can have a time to live ttl it can be correlated within the given ttl after the ttl is reached the message expires and is removed currently the java and the go client use a different default value for the message ttl in the java client the default ttl is in the go client the default ttl is we should align the default value in all official clients to align the general behavior i recommend setting the default message ttl to zero i e disable the message buffering if the ttl is set to a high value then it can impact the performance of the broker see instead of setting an arbitrary value for the ttl let the user explicitly choose to use the buffering and think about a reasonable value | 1 |
306 | 6,396,443,449 | IssuesEvent | 2017-08-04 15:33:40 | rook/rook | https://api.github.com/repos/rook/rook | closed | Cluster fails to come up after a reboot | rados reliability | The mons are dependent on a constant network identity. If that identity changes, nobody will be able to connect to the mon. The mon will be rejected.
Ceph enforces that the mon endpoint is what the mon binds to. Putting a k8s service in front of a mon to have a consistent endpoint does not work.
Currently the mons use the PodIP, which will change at every restart. We must either find a way to soften the ceph endpoint requirements to allow the service translation, or find a way to get a consistent endpoint for everyone to talk to the mons, even after reboot. | True | Cluster fails to come up after a reboot - The mons are dependent on a constant network identity. If that identity changes, nobody will be able to connect to the mon. The mon will be rejected.
Ceph enforces that the mon endpoint is what the mon binds to. Putting a k8s service in front of a mon to have a consistent endpoint does not work.
Currently the mons use the PodIP, which will change at every restart. We must either find a way to soften the ceph endpoint requirements to allow the service translation, or find a way to get a consistent endpoint for everyone to talk to the mons, even after reboot. | reli | cluster fails to come up after a reboot the mons are dependent on a constant network identity if that identity changes nobody will be able to connect to the mon the mon will be rejected ceph enforces that the mon endpoint is what the mon binds to putting a service in front of a mon to have a consistent endpoint does not work currently the mons use the podip which will change at every restart we must either find a way to soften the ceph endpoint requirements to allow the service translation or find a way to get a consistent endpoint for everyone to talk to the mons even after reboot | 1 |
156,814 | 5,989,608,736 | IssuesEvent | 2017-06-02 09:04:52 | universAAL/ui | https://api.github.com/repos/universAAL/ui | closed | UI frozen | bug imported priority 3 | _Originally Opened: @dimokas (2013-03-22 16:02:30_)
_Originally Closed: 2013-04-22 10:56:24_
In some cases, when the service call takes some time (few seconds or less) then the UI is not repsonding. The user cannot press any button since there is no interaction. It is like a frozen window.
--
From: _this issue has been automatically imported from our old issue tracker_ | 1.0 | UI frozen - _Originally Opened: @dimokas (2013-03-22 16:02:30_)
_Originally Closed: 2013-04-22 10:56:24_
In some cases, when the service call takes some time (few seconds or less) then the UI is not repsonding. The user cannot press any button since there is no interaction. It is like a frozen window.
--
From: _this issue has been automatically imported from our old issue tracker_ | non_reli | ui frozen originally opened dimokas originally closed in some cases when the service call takes some time few seconds or less then the ui is not repsonding the user cannot press any button since there is no interaction it is like a frozen window from this issue has been automatically imported from our old issue tracker | 0 |
120,573 | 4,791,766,652 | IssuesEvent | 2016-10-31 13:44:07 | vidalborromeo/eecshelp-beta | https://api.github.com/repos/vidalborromeo/eecshelp-beta | opened | For Moderator Signed-In On Multiple Browsers, First Sign-Off Makes Them Appear "Offline" | Component: Backend Priority: Low Type: Bug | If a moderator is signed in via multiple browsers, the first Unsubscribe event will make it appear that they are no longer online.
Solution: During the backend query in Unsubscribe, check if there are other Observers associated with that User entity. | 1.0 | For Moderator Signed-In On Multiple Browsers, First Sign-Off Makes Them Appear "Offline" - If a moderator is signed in via multiple browsers, the first Unsubscribe event will make it appear that they are no longer online.
Solution: During the backend query in Unsubscribe, check if there are other Observers associated with that User entity. | non_reli | for moderator signed in on multiple browsers first sign off makes them appear offline if a moderator is signed in via multiple browsers the first unsubscribe event will make it appear that they are no longer online solution during the backend query in unsubscribe check if there are other observers associated with that user entity | 0 |
352,589 | 32,078,556,914 | IssuesEvent | 2023-09-25 12:37:39 | durof/status | https://api.github.com/repos/durof/status | closed | 🛑 Mention Test Bot is down | status mention-test-bot | In [`00c5ad2`](https://github.com/durof/status/commit/00c5ad28047456252721048b7f545d3ffa7c9799
), Mention Test Bot (https://backend.isbotdown.com/bots/mentiontestbot) was **down**:
- HTTP code: 200
- Response time: 88 ms
| 1.0 | 🛑 Mention Test Bot is down - In [`00c5ad2`](https://github.com/durof/status/commit/00c5ad28047456252721048b7f545d3ffa7c9799
), Mention Test Bot (https://backend.isbotdown.com/bots/mentiontestbot) was **down**:
- HTTP code: 200
- Response time: 88 ms
| non_reli | 🛑 mention test bot is down in mention test bot was down http code response time ms | 0 |
1,077 | 12,885,581,771 | IssuesEvent | 2020-07-13 08:00:25 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | [BUG] Interminent and recurring JsonParseException | Compute - VM Mgmt customer-reported needs-author-feedback no-recent-activity tenet-reliability | **Describe the bug**
Often when using the sdk or even plain http requests to fetch information from azure, an unexpected response is given.
The response looks like html code, rather then JSON.
***Exception or Stack Trace***
`java.lang.RuntimeException: com.fasterxml.jackson.core.JsonParseException: Unexpected character ('<' (code 60)): expected a valid value (number, String, array, object, 'true', 'false' or 'null')
at [Source: (String)"<!doctype html>
<html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Symantec CloudSOC™</title>
<meta name="robots" content="noindex, nofollow" />
<meta name="description" content="">
<meta http-equiv="Content-Security-Policy" content="default-src * data: blob:;script-src *.elastica.net *.elastica-inc.com *.nr-data.net *.newrelic.com 'unsafe-inline' 'unsafe-eval' blob: data: 'self';style-src data: blob: 'unsafe-inline'"[truncated 1533 chars]; line: 1, column: 2]
at rx.exceptions.Exceptions.propagate(Exceptions.java:57)
at rx.observables.BlockingObservable.forEach(BlockingObservable.java:139) `
**To Reproduce**
Steps to reproduce the behavior:
I am not sure but to me, I just need to execute any sdk list operation.
***Code Snippet***
azure.virtualMachines().listByResouceGroup
And many others.
**Expected behavior**
Every response should be a valid json.
**Setup (please complete the following information):**
- OS: windows
- IDE : intelliJ
- Version of the Library used 1.31.1
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [ x] Bug Description Added
- [ x] Repro Steps Added
- [x ] Setup information Added
| True | [BUG] Interminent and recurring JsonParseException - **Describe the bug**
Often when using the sdk or even plain http requests to fetch information from azure, an unexpected response is given.
The response looks like html code, rather then JSON.
***Exception or Stack Trace***
`java.lang.RuntimeException: com.fasterxml.jackson.core.JsonParseException: Unexpected character ('<' (code 60)): expected a valid value (number, String, array, object, 'true', 'false' or 'null')
at [Source: (String)"<!doctype html>
<html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Symantec CloudSOC™</title>
<meta name="robots" content="noindex, nofollow" />
<meta name="description" content="">
<meta http-equiv="Content-Security-Policy" content="default-src * data: blob:;script-src *.elastica.net *.elastica-inc.com *.nr-data.net *.newrelic.com 'unsafe-inline' 'unsafe-eval' blob: data: 'self';style-src data: blob: 'unsafe-inline'"[truncated 1533 chars]; line: 1, column: 2]
at rx.exceptions.Exceptions.propagate(Exceptions.java:57)
at rx.observables.BlockingObservable.forEach(BlockingObservable.java:139) `
**To Reproduce**
Steps to reproduce the behavior:
I am not sure but to me, I just need to execute any sdk list operation.
***Code Snippet***
azure.virtualMachines().listByResouceGroup
And many others.
**Expected behavior**
Every response should be a valid json.
**Setup (please complete the following information):**
- OS: windows
- IDE : intelliJ
- Version of the Library used 1.31.1
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [ x] Bug Description Added
- [ x] Repro Steps Added
- [x ] Setup information Added
| reli | interminent and recurring jsonparseexception describe the bug often when using the sdk or even plain http requests to fetch information from azure an unexpected response is given the response looks like html code rather then json exception or stack trace java lang runtimeexception com fasterxml jackson core jsonparseexception unexpected character code expected a valid value number string array object true false or null at source string symantec cloudsoc™ meta http equiv content security policy content default src data blob script src elastica net elastica inc com nr data net newrelic com unsafe inline unsafe eval blob data self style src data blob unsafe inline line column at rx exceptions exceptions propagate exceptions java at rx observables blockingobservable foreach blockingobservable java to reproduce steps to reproduce the behavior i am not sure but to me i just need to execute any sdk list operation code snippet azure virtualmachines listbyresoucegroup and many others expected behavior every response should be a valid json setup please complete the following information os windows ide intellij version of the library used information checklist kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report bug description added repro steps added setup information added | 1 |
389,614 | 26,825,331,072 | IssuesEvent | 2023-02-02 12:28:20 | nextauthjs/next-auth | https://api.github.com/repos/nextauthjs/next-auth | closed | Fix prisma doc schema.prisma userId warning for mongodb | documentation triage | ### What is the improvement or update you wish to see?
Don't show the warning message.
**"Warning validating field user in model Account: Field Account.userId and User.id must have the same native type for MongoDB to join those collections correctly. Consider either removing User.id's native type attribute or adding '@db.ObjectId' to Account.userId. Beware that this will become an error in the future."**
### Is there any context that might help us understand?
Prisma member recommended add `@db.ObjectId` for `userId` field to fix this problem. [link](https://github.com/prisma/prisma/discussions/17624#discussioncomment-4820223)
### Does the docs page already exist? Please link to it.
https://next-auth.js.org/adapters/prisma | 1.0 | Fix prisma doc schema.prisma userId warning for mongodb - ### What is the improvement or update you wish to see?
Don't show the warning message.
**"Warning validating field user in model Account: Field Account.userId and User.id must have the same native type for MongoDB to join those collections correctly. Consider either removing User.id's native type attribute or adding '@db.ObjectId' to Account.userId. Beware that this will become an error in the future."**
### Is there any context that might help us understand?
Prisma member recommended add `@db.ObjectId` for `userId` field to fix this problem. [link](https://github.com/prisma/prisma/discussions/17624#discussioncomment-4820223)
### Does the docs page already exist? Please link to it.
https://next-auth.js.org/adapters/prisma | non_reli | fix prisma doc schema prisma userid warning for mongodb what is the improvement or update you wish to see don t show the warning message warning validating field user in model account field account userid and user id must have the same native type for mongodb to join those collections correctly consider either removing user id s native type attribute or adding db objectid to account userid beware that this will become an error in the future is there any context that might help us understand prisma member recommended add db objectid for userid field to fix this problem does the docs page already exist please link to it | 0 |
303,274 | 22,961,863,212 | IssuesEvent | 2022-07-19 16:00:15 | Workday/canvas-kit | https://api.github.com/repos/Workday/canvas-kit | closed | Update ActionBar section in V7 upgrade guide | documentation s:1 7.x | ## 💡 Suggestion
The ActionBar section in the V7 upgrade guide could use a bit more detail on the removal of the `fixed` prop.
## Motivation
There have been a few support questions that have come up recently. | 1.0 | Update ActionBar section in V7 upgrade guide - ## 💡 Suggestion
The ActionBar section in the V7 upgrade guide could use a bit more detail on the removal of the `fixed` prop.
## Motivation
There have been a few support questions that have come up recently. | non_reli | update actionbar section in upgrade guide 💡 suggestion the actionbar section in the upgrade guide could use a bit more detail on the removal of the fixed prop motivation there have been a few support questions that have come up recently | 0 |
46,631 | 13,055,950,338 | IssuesEvent | 2020-07-30 03:12:43 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | [docs] Broken on combo trunk (Trac #1613) | Incomplete Migration Migrated from Trac cmake defect | Migrated from https://code.icecube.wisc.edu/ticket/1613
```json
{
"status": "closed",
"changetime": "2016-04-05T23:31:03",
"description": "This is the error I get on combo trunk Ubuntu 14.04 with sphinx 1.4 (recently updated 'cause the napolean extension is now required).\n\nException occurred:\n File \"/usr/local/lib/python2.7/dist-packages/sphinx/domains/cpp.py\", line 4039, in _resolve_xref_inner\n assert parentSymbol # should be there\nAssertionError\n",
"reporter": "olivas",
"cc": "",
"resolution": "worksforme",
"_ts": "1459899063339623",
"component": "cmake",
"summary": "[docs] Broken on combo trunk",
"priority": "major",
"keywords": "",
"time": "2016-03-30T13:43:09",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | [docs] Broken on combo trunk (Trac #1613) - Migrated from https://code.icecube.wisc.edu/ticket/1613
```json
{
"status": "closed",
"changetime": "2016-04-05T23:31:03",
"description": "This is the error I get on combo trunk Ubuntu 14.04 with sphinx 1.4 (recently updated 'cause the napolean extension is now required).\n\nException occurred:\n File \"/usr/local/lib/python2.7/dist-packages/sphinx/domains/cpp.py\", line 4039, in _resolve_xref_inner\n assert parentSymbol # should be there\nAssertionError\n",
"reporter": "olivas",
"cc": "",
"resolution": "worksforme",
"_ts": "1459899063339623",
"component": "cmake",
"summary": "[docs] Broken on combo trunk",
"priority": "major",
"keywords": "",
"time": "2016-03-30T13:43:09",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| non_reli | broken on combo trunk trac migrated from json status closed changetime description this is the error i get on combo trunk ubuntu with sphinx recently updated cause the napolean extension is now required n nexception occurred n file usr local lib dist packages sphinx domains cpp py line in resolve xref inner n assert parentsymbol should be there nassertionerror n reporter olivas cc resolution worksforme ts component cmake summary broken on combo trunk priority major keywords time milestone owner nega type defect | 0 |
697 | 9,934,215,224 | IssuesEvent | 2019-07-02 14:02:09 | plotly/dash-table | https://api.github.com/repos/plotly/dash-table | closed | add test for semi editable table | Attribute: Reliability | verify setting editable=True at table level and editable=False at column level to disable editing of specified columns works. | True | add test for semi editable table - verify setting editable=True at table level and editable=False at column level to disable editing of specified columns works. | reli | add test for semi editable table verify setting editable true at table level and editable false at column level to disable editing of specified columns works | 1 |
780,308 | 27,389,105,401 | IssuesEvent | 2023-02-28 15:11:07 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | GOCA: Vm migration error -501 | Status: Pending Type: Bug Category: API Priority: Normal | OpenNebula 6.0.0.2
github.com/OpenNebula/one/src/oca/go/src/goca v0.0.0-20230130124959-3c9bfecbd197
Try
if err := n.controller.VM(134).Migrate(1, true, true, 0, 0); err != nil {
return err
}
but have err:
Code: -501
String: "Parameter that is supposed to be integer is not"
| 1.0 | GOCA: Vm migration error -501 - OpenNebula 6.0.0.2
github.com/OpenNebula/one/src/oca/go/src/goca v0.0.0-20230130124959-3c9bfecbd197
Try
if err := n.controller.VM(134).Migrate(1, true, true, 0, 0); err != nil {
return err
}
but have err:
Code: -501
String: "Parameter that is supposed to be integer is not"
| non_reli | goca vm migration error opennebula github com opennebula one src oca go src goca try if err n controller vm migrate true true err nil return err but have err code string parameter that is supposed to be integer is not | 0 |
1,185 | 13,634,351,343 | IssuesEvent | 2020-09-24 23:35:50 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | opened | Ensure all Key Vault models align correctly with the Swagger specification | Client KeyVault tenet-reliability | Issue #14513 showed that the Key Vault Keys library had issues when deserializing a couple models (fixed by #15375), so it would be a worthwhile effort to further look into the rest of the models in Keys, Certificates and Secrets to ensure everything conforms to the service's Swagger specification. | True | Ensure all Key Vault models align correctly with the Swagger specification - Issue #14513 showed that the Key Vault Keys library had issues when deserializing a couple models (fixed by #15375), so it would be a worthwhile effort to further look into the rest of the models in Keys, Certificates and Secrets to ensure everything conforms to the service's Swagger specification. | reli | ensure all key vault models align correctly with the swagger specification issue showed that the key vault keys library had issues when deserializing a couple models fixed by so it would be a worthwhile effort to further look into the rest of the models in keys certificates and secrets to ensure everything conforms to the service s swagger specification | 1 |
77 | 3,477,830,527 | IssuesEvent | 2015-12-28 06:21:47 | PopupMaker/Popup-Maker | https://api.github.com/repos/PopupMaker/Popup-Maker | closed | Introduce new Popup class. | Reliability | This will simplify the data model & methods needed for both standard template loading and WP API loading.
Methods
is_loadable()
title()
content()
theme_id()
close_text() | True | Introduce new Popup class. - This will simplify the data model & methods needed for both standard template loading and WP API loading.
Methods
is_loadable()
title()
content()
theme_id()
close_text() | reli | introduce new popup class this will simplify the data model methods needed for both standard template loading and wp api loading methods is loadable title content theme id close text | 1 |
278,999 | 21,104,597,275 | IssuesEvent | 2022-04-04 17:27:33 | WaffleKone/Blades-Companion | https://api.github.com/repos/WaffleKone/Blades-Companion | opened | Configure boiler plates and linting | documentation | Make it look pretty and good!
- [ ] Set up and configure ESLint and Prettier and make them work together
- [ ] Set up automatic testing and changes with ESLint when commits are made
- [ ] Set up all dependencies in package.json
| 1.0 | Configure boiler plates and linting - Make it look pretty and good!
- [ ] Set up and configure ESLint and Prettier and make them work together
- [ ] Set up automatic testing and changes with ESLint when commits are made
- [ ] Set up all dependencies in package.json
| non_reli | configure boiler plates and linting make it look pretty and good set up and configure eslint and prettier and make them work together set up automatic testing and changes with eslint when commits are made set up all dependencies in package json | 0 |
300 | 6,150,716,165 | IssuesEvent | 2017-06-27 23:31:29 | CCI-MOC/hil | https://api.github.com/repos/CCI-MOC/hil | closed | Using SNMP to control switches to set VLANS | Reliability | We should have a better way to set VLANS programmatically than using pexpect due to the problems discussed in #744. @pjd-nu gave us super overview of how SNMP works and what troubles we may encounter (like it is good for reading stuff from hardware, but writing to hardware can be tedious at times)
I, @henn, @naved001, @pjd-nu agree that this can be a good internship project.
Explore, understand SNMP and come up with a solution to set VLANS using it. | True | Using SNMP to control switches to set VLANS - We should have a better way to set VLANS programmatically than using pexpect due to the problems discussed in #744. @pjd-nu gave us super overview of how SNMP works and what troubles we may encounter (like it is good for reading stuff from hardware, but writing to hardware can be tedious at times)
I, @henn, @naved001, @pjd-nu agree that this can be a good internship project.
Explore, understand SNMP and come up with a solution to set VLANS using it. | reli | using snmp to control switches to set vlans we should have a better way to set vlans programmatically than using pexpect due to the problems discussed in pjd nu gave us super overview of how snmp works and what troubles we may encounter like it is good for reading stuff from hardware but writing to hardware can be tedious at times i henn pjd nu agree that this can be a good internship project explore understand snmp and come up with a solution to set vlans using it | 1 |
19,684 | 13,389,959,492 | IssuesEvent | 2020-09-02 19:44:59 | revangel/nostalgiabot2 | https://api.github.com/repos/revangel/nostalgiabot2 | opened | Add automatic linting | Technical infrastructure | # Overview
It'd be nice to have our code consistently formatted, and more importantly, automatically. Use a combination of flake-8 and [black](https://github.com/psf/black) to implement this
# Acceptance Criteria
- [ ] A pre-commit hook exists to automatically format code changes | 1.0 | Add automatic linting - # Overview
It'd be nice to have our code consistently formatted, and more importantly, automatically. Use a combination of flake-8 and [black](https://github.com/psf/black) to implement this
# Acceptance Criteria
- [ ] A pre-commit hook exists to automatically format code changes | non_reli | add automatic linting overview it d be nice to have our code consistently formatted and more importantly automatically use a combination of flake and to implement this acceptance criteria a pre commit hook exists to automatically format code changes | 0 |
588 | 8,743,752,664 | IssuesEvent | 2018-12-12 20:05:24 | rook/rook | https://api.github.com/repos/rook/rook | closed | Support user-space driver | block perf reliability wontfix | <!-- **Are you in the right place?**
1. For issues or feature requests, please create an issue in this repository.
2. For general technical and non-technical questions, we are happy to help you on our [Rook.io Slack](https://Rook-io.slack.com).
3. Did you already search the existing open issues for anything similar? -->
**Is this a bug report or feature request?**
* Feature Request
**What should the feature do:**
I have tried to use `rbd-nbd` feature-rich client to support ceph rbd image. But I noticed it was restricted by existing mechanisms. When using rbd-nbd to map rbd image, it would create a user-space process to deal with network IO. Currently ceph client map rbd image in the agent pod. If I use `rbd-nbd` client in the same way, I found that
1. `rbd-nbd list-mapped` not work. Because the rbd-nbd process have different process id in the container and the host.
2. the rook agent became not stateless. The rbd-nbd process would prevent agent stop/restar.
I think any user-space driver like `rbd-nbd` would face similar problems. Because the key reason is that the driver process is running in the container but not the host. I hope we can build an unified architecture to support this type driver. It is more suitable that any driver that use user-space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly.
**What is use case behind this feature:**
**Environment**:
<!-- Specific environment information that helps with the feature request -->
| True | Support user-space driver - <!-- **Are you in the right place?**
1. For issues or feature requests, please create an issue in this repository.
2. For general technical and non-technical questions, we are happy to help you on our [Rook.io Slack](https://Rook-io.slack.com).
3. Did you already search the existing open issues for anything similar? -->
**Is this a bug report or feature request?**
* Feature Request
**What should the feature do:**
I have tried to use `rbd-nbd` feature-rich client to support ceph rbd image. But I noticed it was restricted by existing mechanisms. When using rbd-nbd to map rbd image, it would create a user-space process to deal with network IO. Currently ceph client map rbd image in the agent pod. If I use `rbd-nbd` client in the same way, I found that
1. `rbd-nbd list-mapped` not work. Because the rbd-nbd process have different process id in the container and the host.
2. the rook agent became not stateless. The rbd-nbd process would prevent agent stop/restar.
I think any user-space driver like `rbd-nbd` would face similar problems. Because the key reason is that the driver process is running in the container but not the host. I hope we can build an unified architecture to support this type driver. It is more suitable that any driver that use user-space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly.
**What is use case behind this feature:**
**Environment**:
<!-- Specific environment information that helps with the feature request -->
| reli | support user space driver are you in the right place for issues or feature requests please create an issue in this repository for general technical and non technical questions we are happy to help you on our did you already search the existing open issues for anything similar is this a bug report or feature request feature request what should the feature do i have tried to use rbd nbd feature rich client to support ceph rbd image but i noticed it was restricted by existing mechanisms when using rbd nbd to map rbd image it would create a user space process to deal with network io currently ceph client map rbd image in the agent pod if i use rbd nbd client in the same way i found that rbd nbd list mapped not work because the rbd nbd process have different process id in the container and the host the rook agent became not stateless the rbd nbd process would prevent agent stop restar i think any user space driver like rbd nbd would face similar problems because the key reason is that the driver process is running in the container but not the host i hope we can build an unified architecture to support this type driver it is more suitable that any driver that use user space process should not be started by the rook agent in a container, but by the flexvolume plugin in the node directly what is use case behind this feature environment | 1 |
2,315 | 24,682,950,719 | IssuesEvent | 2022-10-18 23:42:26 | microsoft/pxt-arcade | https://api.github.com/repos/microsoft/pxt-arcade | closed | "maximum file size in package is ~64k" error causes game crash | bug p2 reliability forum | See: https://forum.makecode.com/t/maximum-file-size-in-package-is-64k/5267/2
I'm asking for more details and will update if we learn more. | True | "maximum file size in package is ~64k" error causes game crash - See: https://forum.makecode.com/t/maximum-file-size-in-package-is-64k/5267/2
I'm asking for more details and will update if we learn more. | reli | maximum file size in package is error causes game crash see i m asking for more details and will update if we learn more | 1 |
2,731 | 27,298,852,026 | IssuesEvent | 2023-02-23 23:05:39 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | opened | [FEA] Run CI tests with injected OOMs | feature request ? - Needs Triage reliability | The idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically. A retried test that passes CI (where the output is the same as the CPU) is very valuable and much better at stressing this code out, with the alternative being replicated unit tests that touch a narrower scope.
I have prototyped this locally, but can't really PR it yet because I need some hooks from `RmmSpark` and thread association/disassociation that @revans2 is plumbing with the retry framework already. I am thinking about this in the context of code using `withRetry` (https://github.com/NVIDIA/spark-rapids/issues/7256). This gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions. Code outside `withRetry` doesn't have a natural entry points that I can think of, and it wouldn't know how to retry.
We need the tests to be very loud about when they are getting an injected OOM. What I have prototyped adds "INJECT_OOM" in the test name, so it should be really easy to figure out that a failure is likely related to OOM handling.
```
../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[32][INJECT_OOM]
```
vs
```
../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[33]
``` | True | [FEA] Run CI tests with injected OOMs - The idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically. A retried test that passes CI (where the output is the same as the CPU) is very valuable and much better at stressing this code out, with the alternative being replicated unit tests that touch a narrower scope.
I have prototyped this locally, but can't really PR it yet because I need some hooks from `RmmSpark` and thread association/disassociation that @revans2 is plumbing with the retry framework already. I am thinking about this in the context of code using `withRetry` (https://github.com/NVIDIA/spark-rapids/issues/7256). This gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions. Code outside `withRetry` doesn't have a natural entry points that I can think of, and it wouldn't know how to retry.
We need the tests to be very loud about when they are getting an injected OOM. What I have prototyped adds "INJECT_OOM" in the test name, so it should be really easy to figure out that a failure is likely related to OOM handling.
```
../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[32][INJECT_OOM]
```
vs
```
../../src/main/python/hash_aggregate_test.py::test_hash_reduction_decimal_overflow_sum[33]
``` | reli | run ci tests with injected ooms the idea behind this task is to leverage our python integration tests to perform retries likely randomly but could be configured to happen deterministically a retried test that passes ci where the output is the same as the cpu is very valuable and much better at stressing this code out with the alternative being replicated unit tests that touch a narrower scope i have prototyped this locally but can t really pr it yet because i need some hooks from rmmspark and thread association disassociation that is plumbing with the retry framework already i am thinking about this in the context of code using withretry this gives us a natural entry for a config to be read that allows us to selectively inject the retry exceptions code outside withretry doesn t have a natural entry points that i can think of and it wouldn t know how to retry we need the tests to be very loud about when they are getting an injected oom what i have prototyped adds inject oom in the test name so it should be really easy to figure out that a failure is likely related to oom handling src main python hash aggregate test py test hash reduction decimal overflow sum vs src main python hash aggregate test py test hash reduction decimal overflow sum | 1 |
733,113 | 25,288,537,765 | IssuesEvent | 2022-11-16 21:34:06 | thoth-station/kebechet | https://api.github.com/repos/thoth-station/kebechet | closed | descriptive text is wrong | kind/bug priority/important-soon sig/user-experience | ## Bug description
The descriptive text of the first bullet point at https://github.com/thoth-station/thamos/pull/1180 is commenting on the wrong object. the text is on a PR, but describes what happens on an issue.
### Steps to Reproduce
Steps to reproduce the behavior:
1. Go to https://github.com/thoth-station/thamos/pull/1180
2. see body of PR
4. See error
### Actual behavior
won't object used in sentence.
### Expected behavior
PR should be the object
### Environment information
n/a
### Additional context
n/a
/priority important-soon
/sig user-experience | 1.0 | descriptive text is wrong - ## Bug description
The descriptive text of the first bullet point at https://github.com/thoth-station/thamos/pull/1180 is commenting on the wrong object. the text is on a PR, but describes what happens on an issue.
### Steps to Reproduce
Steps to reproduce the behavior:
1. Go to https://github.com/thoth-station/thamos/pull/1180
2. see body of PR
4. See error
### Actual behavior
won't object used in sentence.
### Expected behavior
PR should be the object
### Environment information
n/a
### Additional context
n/a
/priority important-soon
/sig user-experience | non_reli | descriptive text is wrong bug description the descriptive text of the first bullet point at is commenting on the wrong object the text is on a pr but describes what happens on an issue steps to reproduce steps to reproduce the behavior go to see body of pr see error actual behavior won t object used in sentence expected behavior pr should be the object environment information n a additional context n a priority important soon sig user experience | 0 |
1,698 | 18,836,732,741 | IssuesEvent | 2021-11-11 02:27:46 | zer0Kerbal/OhScrap | https://api.github.com/repos/zer0Kerbal/OhScrap | opened | Oh Scrap!/Scrapyard support? | mod support reliability | From Kerbalism created by [baldamundo](https://github.com/baldamundo): Kerbalism/Kerbalism#493
Wondered if anyone has given this any thought?
Oh Scrap (https://forum.kerbalspaceprogram.com/index.php?/topic/160854-17x-oh-scrap-a-scrapyard-based-part-failure-and-reliability-mod-163-22062019/) is an incredibly good part failure mod, which ties the failure to how many times you've built copies of that part, as well as how often that particular instance of the part has been re-used. This gives a huge amount of depth to a career game, encouraging using common hardware across your stable of vehicles, giving an actual reason for test-flights, significantly nerfing/balancing reuseable vehicles, etc.
At present, the two mods seem to be perfectly compatible, and actually seem to complement each other quite nicely since the Oh Scrap failures only happen while a vessel is loaded - so generally concentrated around launch failures - while Kerbalism's happen in the background and thus better representing long-term decay.
Ultimately the Oh Scrap system is much deeper than Kerbalism's, but Kerbalism has two advantages:
1. Part failures in Kerbalism can occur during background simulation.
2. Failed parts in Kerbalism are displayed in the Kerbalism UI.
Is it at all feasible to integrate the two systems better? I'm guessing that either adding background simulation to Oh Scrap or part history tracking to Kerbalism would be very much non-trivial, but at a minimum, is there a way for Oh Scrap to plug into Kerbalism's API and show its failures in the Kerbalism UI? | True | Oh Scrap!/Scrapyard support? - From Kerbalism created by [baldamundo](https://github.com/baldamundo): Kerbalism/Kerbalism#493
Wondered if anyone has given this any thought?
Oh Scrap (https://forum.kerbalspaceprogram.com/index.php?/topic/160854-17x-oh-scrap-a-scrapyard-based-part-failure-and-reliability-mod-163-22062019/) is an incredibly good part failure mod, which ties the failure to how many times you've built copies of that part, as well as how often that particular instance of the part has been re-used. This gives a huge amount of depth to a career game, encouraging using common hardware across your stable of vehicles, giving an actual reason for test-flights, significantly nerfing/balancing reuseable vehicles, etc.
At present, the two mods seem to be perfectly compatible, and actually seem to complement each other quite nicely since the Oh Scrap failures only happen while a vessel is loaded - so generally concentrated around launch failures - while Kerbalism's happen in the background and thus better representing long-term decay.
Ultimately the Oh Scrap system is much deeper than Kerbalism's, but Kerbalism has two advantages:
1. Part failures in Kerbalism can occur during background simulation.
2. Failed parts in Kerbalism are displayed in the Kerbalism UI.
Is it at all feasible to integrate the two systems better? I'm guessing that either adding background simulation to Oh Scrap or part history tracking to Kerbalism would be very much non-trivial, but at a minimum, is there a way for Oh Scrap to plug into Kerbalism's API and show its failures in the Kerbalism UI? | reli | oh scrap scrapyard support from kerbalism created by kerbalism kerbalism wondered if anyone has given this any thought oh scrap is an incredibly good part failure mod which ties the failure to how many times you ve built copies of that part as well as how often that particular instance of the part has been re used this gives a huge amount of depth to a career game encouraging using common hardware across your stable of vehicles giving an actual reason for test flights significantly nerfing balancing reuseable vehicles etc at present the two mods seem to be perfectly compatible and actually seem to complement each other quite nicely since the oh scrap failures only happen while a vessel is loaded so generally concentrated around launch failures while kerbalism s happen in the background and thus better representing long term decay ultimately the oh scrap system is much deeper than kerbalism s but kerbalism has two advantages part failures in kerbalism can occur during background simulation failed parts in kerbalism are displayed in the kerbalism ui is it at all feasible to integrate the two systems better i m guessing that either adding background simulation to oh scrap or part history tracking to kerbalism would be very much non trivial but at a minimum is there a way for oh scrap to plug into kerbalism s api and show its failures in the kerbalism ui | 1 |
125,845 | 17,861,250,044 | IssuesEvent | 2021-09-06 01:00:06 | maorkuriel/JavaVulnerableLab | https://api.github.com/repos/maorkuriel/JavaVulnerableLab | opened | CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar | security vulnerability | ## CVE-2015-4852 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to dependency file: JavaVulnerableLab/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar,JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- hibernate-core-4.0.1.Final.jar (Root Library)
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.
<p>Publish Date: 2015-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p>
<p>Release Date: 2015-11-18</p>
<p>Fix Resolution: commons-collections:commons-collections:3.2.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.hibernate:hibernate-core:4.0.1.Final;commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-collections:commons-collections:3.2.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-4852","vulnerabilityDetails":"The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> --> | True | CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-4852 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to dependency file: JavaVulnerableLab/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar,JavaVulnerableLab/WEB-INF/lib/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- hibernate-core-4.0.1.Final.jar (Root Library)
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.
<p>Publish Date: 2015-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p>
<p>Release Date: 2015-11-18</p>
<p>Fix Resolution: commons-collections:commons-collections:3.2.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-collections","packageName":"commons-collections","packageVersion":"3.2.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.hibernate:hibernate-core:4.0.1.Final;commons-collections:commons-collections:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-collections:commons-collections:3.2.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-4852","vulnerabilityDetails":"The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> --> | non_reli | cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file javavulnerablelab pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar javavulnerablelab web inf lib commons collections jar dependency hierarchy hibernate core final jar root library x commons collections jar vulnerable library found in base branch master vulnerability details the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org hibernate hibernate core final commons collections commons collections isminimumfixversionavailable true minimumfixversion commons collections commons collections basebranches vulnerabilityidentifier cve vulnerabilitydetails the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product vulnerabilityurl | 0 |
2,761 | 27,557,522,324 | IssuesEvent | 2023-03-07 19:08:28 | rapidsai/cudf | https://api.github.com/repos/rapidsai/cudf | opened | [FEA][JNI] Leverage cub's multi-buffer copy algorithm in JNI bindings | feature request ? - Needs Triage Performance cuDF (Java) helps: Spark Reliability | We would like to use https://github.com/NVIDIA/cub/issues/297 for our spilling logic https://github.com/NVIDIA/spark-rapids/issues/7672 in the spark-rapids plugin.
We currently contiguous_split every single buffer that _could be_ spilled in the future, because at the time of spill the last thing we want to do is double our memory usage right before copying to host. The multi-buffer work should allow us to bounce-buffer out of GPU memory at spill time, without having to contiguous_split for spillable buffers. | True | [FEA][JNI] Leverage cub's multi-buffer copy algorithm in JNI bindings - We would like to use https://github.com/NVIDIA/cub/issues/297 for our spilling logic https://github.com/NVIDIA/spark-rapids/issues/7672 in the spark-rapids plugin.
We currently contiguous_split every single buffer that _could be_ spilled in the future, because at the time of spill the last thing we want to do is double our memory usage right before copying to host. The multi-buffer work should allow us to bounce-buffer out of GPU memory at spill time, without having to contiguous_split for spillable buffers. | reli | leverage cub s multi buffer copy algorithm in jni bindings we would like to use for our spilling logic in the spark rapids plugin we currently contiguous split every single buffer that could be spilled in the future because at the time of spill the last thing we want to do is double our memory usage right before copying to host the multi buffer work should allow us to bounce buffer out of gpu memory at spill time without having to contiguous split for spillable buffers | 1 |
50,032 | 7,551,601,394 | IssuesEvent | 2018-04-18 20:40:23 | kubernetes/kubeadm | https://api.github.com/repos/kubernetes/kubeadm | closed | Please document that self-hosted feature does not survive a reboot (v 1.9.3) | active kind/documentation kind/feature priority/important-soon | <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.-->
## Is this a BUG REPORT or FEATURE REQUEST?
FEATURE REQUEST
## Versions
**kubeadm version** (use `kubeadm version`):
```
core@localhost ~ $ kubeadm version
kubeadm version: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
```
**Environment**:
- **Kubernetes version** (use `kubectl version`):
```
core@localhost ~ $ kubectl version
Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T12:22:21Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
```
- **Cloud provider or hardware configuration**:
Dell R720 Bare Metal Server
- **OS** (e.g. from /etc/os-release):
```
core@localhost ~ $ cat /etc/os-release
NAME="Container Linux by CoreOS"
ID=coreos
VERSION=1632.3.0
VERSION_ID=1632.3.0
BUILD_ID=2018-02-14-0338
PRETTY_NAME="Container Linux by CoreOS 1632.3.0 (Ladybug)"
ANSI_COLOR="38;5;75"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://issues.coreos.com"
COREOS_BOARD="amd64-usr"
```
- **Kernel** (e.g. `uname -a`):
```
core@localhost ~ $ uname -a
Linux localhost 4.14.19-coreos #1 SMP Wed Feb 14 03:18:05 UTC 2018 x86_64 Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz GenuineIntel GNU/Linux
```
- **Others**:
## What happened?
Please document that the self-hosted feature does not work when you reboot the server without additional configuration. Kubernetes cluster does not restart after reboot
## What you expected to happen?
Expected it to restart.
## How to reproduce it (as minimally and precisely as possible)?
```
sudo kubeadm init --feature-gates=SelfHosting=true --pod-network-cidr=10.244.0.0/16
sudo systemctl enable kubelet.service
sudo shutdown -r now
```
## Anything else we need to know?
A workaround is described here: https://github.com/xetys/k8s-self-hosted-recovery
It took me a long time to find it because I didn't understand what was going wrong so it was hard to search. Once I understood that the API service wasn't starting and it was a catch-22 problem due to the self-hosted feature, then it was pretty each to search and find the workaround.
| 1.0 | Please document that self-hosted feature does not survive a reboot (v 1.9.3) - <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.-->
## Is this a BUG REPORT or FEATURE REQUEST?
FEATURE REQUEST
## Versions
**kubeadm version** (use `kubeadm version`):
```
core@localhost ~ $ kubeadm version
kubeadm version: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
```
**Environment**:
- **Kubernetes version** (use `kubectl version`):
```
core@localhost ~ $ kubectl version
Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T12:22:21Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.3", GitCommit:"d2835416544f298c919e2ead3be3d0864b52323b", GitTreeState:"clean", BuildDate:"2018-02-07T11:55:20Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
```
- **Cloud provider or hardware configuration**:
Dell R720 Bare Metal Server
- **OS** (e.g. from /etc/os-release):
```
core@localhost ~ $ cat /etc/os-release
NAME="Container Linux by CoreOS"
ID=coreos
VERSION=1632.3.0
VERSION_ID=1632.3.0
BUILD_ID=2018-02-14-0338
PRETTY_NAME="Container Linux by CoreOS 1632.3.0 (Ladybug)"
ANSI_COLOR="38;5;75"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://issues.coreos.com"
COREOS_BOARD="amd64-usr"
```
- **Kernel** (e.g. `uname -a`):
```
core@localhost ~ $ uname -a
Linux localhost 4.14.19-coreos #1 SMP Wed Feb 14 03:18:05 UTC 2018 x86_64 Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz GenuineIntel GNU/Linux
```
- **Others**:
## What happened?
Please document that the self-hosted feature does not work when you reboot the server without additional configuration. Kubernetes cluster does not restart after reboot
## What you expected to happen?
Expected it to restart.
## How to reproduce it (as minimally and precisely as possible)?
```
sudo kubeadm init --feature-gates=SelfHosting=true --pod-network-cidr=10.244.0.0/16
sudo systemctl enable kubelet.service
sudo shutdown -r now
```
## Anything else we need to know?
A workaround is described here: https://github.com/xetys/k8s-self-hosted-recovery
It took me a long time to find it because I didn't understand what was going wrong so it was hard to search. Once I understood that the API service wasn't starting and it was a catch-22 problem due to the self-hosted feature, then it was pretty each to search and find the workaround.
| non_reli | please document that self hosted feature does not survive a reboot v is this a bug report or feature request feature request versions kubeadm version use kubeadm version core localhost kubeadm version kubeadm version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux environment kubernetes version use kubectl version core localhost kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux cloud provider or hardware configuration dell bare metal server os e g from etc os release core localhost cat etc os release name container linux by coreos id coreos version version id build id pretty name container linux by coreos ladybug ansi color home url bug report url coreos board usr kernel e g uname a core localhost uname a linux localhost coreos smp wed feb utc intel r xeon r cpu genuineintel gnu linux others what happened please document that the self hosted feature does not work when you reboot the server without additional configuration kubernetes cluster does not restart after reboot what you expected to happen expected it to restart how to reproduce it as minimally and precisely as possible sudo kubeadm init feature gates selfhosting true pod network cidr sudo systemctl enable kubelet service sudo shutdown r now anything else we need to know a workaround is described here it took me a long time to find it because i didn t understand what was going wrong so it was hard to search once i understood that the api service wasn t starting and it was a catch problem due to the self hosted feature then it was pretty each to search and find the workaround | 0 |
29,629 | 24,122,163,495 | IssuesEvent | 2022-09-20 19:46:36 | GCTC-NTGC/gc-digital-talent | https://api.github.com/repos/GCTC-NTGC/gc-digital-talent | closed | Not possible to remove a feature flag on a single vertical | infrastructure | When removing a feature flag it would be good to be able to do it progressively: dev, UAT, and finally prod. Right now the setting is defined in appsvc-VERTICAL-talentcloud-template.json which is the shared file. So it isn't possible, I think. | 1.0 | Not possible to remove a feature flag on a single vertical - When removing a feature flag it would be good to be able to do it progressively: dev, UAT, and finally prod. Right now the setting is defined in appsvc-VERTICAL-talentcloud-template.json which is the shared file. So it isn't possible, I think. | non_reli | not possible to remove a feature flag on a single vertical when removing a feature flag it would be good to be able to do it progressively dev uat and finally prod right now the setting is defined in appsvc vertical talentcloud template json which is the shared file so it isn t possible i think | 0 |
1,932 | 21,897,270,176 | IssuesEvent | 2022-05-20 09:50:28 | jina-ai/jina | https://api.github.com/repos/jina-ai/jina | closed | Smarter retry mechanism for Executors with replicas | epic/reliability | Currently, the round robin scheduling between replicas does not apply when retrying to send a request, but only when a new request comes in. This means that if a replica is down, the gateway stubbornly tries to connect to the this same replica, ultimately failing the request despite other replicas being available.
This should be changed s.t. round robing scheduling is applied also to connection retries. | True | Smarter retry mechanism for Executors with replicas - Currently, the round robin scheduling between replicas does not apply when retrying to send a request, but only when a new request comes in. This means that if a replica is down, the gateway stubbornly tries to connect to the this same replica, ultimately failing the request despite other replicas being available.
This should be changed s.t. round robing scheduling is applied also to connection retries. | reli | smarter retry mechanism for executors with replicas currently the round robin scheduling between replicas does not apply when retrying to send a request but only when a new request comes in this means that if a replica is down the gateway stubbornly tries to connect to the this same replica ultimately failing the request despite other replicas being available this should be changed s t round robing scheduling is applied also to connection retries | 1 |
757,487 | 26,514,714,957 | IssuesEvent | 2023-01-18 19:49:56 | ramp4-pcar4/storylines-editor | https://api.github.com/repos/ramp4-pcar4/storylines-editor | closed | Editor slides do not switch properly when switching between panels of the same type | Bug Priority: High Needs: estimate | If you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type, the contents of the page are not updated properly.
This is an issue we had with dynamic panels in the main Storylines app too, has something to do with Vue not re-rendering everything.
Steps to Reproduce:
1. Click on the `Oil sands extraction` slide and switch to the right panel.
2. Click on the `In-situ extraction` slide.
3. Notice that the image displayed is still for `Oil Sands Extraction`.
*Demo*
 | 1.0 | Editor slides do not switch properly when switching between panels of the same type - If you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type, the contents of the page are not updated properly.
This is an issue we had with dynamic panels in the main Storylines app too, has something to do with Vue not re-rendering everything.
Steps to Reproduce:
1. Click on the `Oil sands extraction` slide and switch to the right panel.
2. Click on the `In-situ extraction` slide.
3. Notice that the image displayed is still for `Oil Sands Extraction`.
*Demo*
 | non_reli | editor slides do not switch properly when switching between panels of the same type if you click on a slide and switch to edit the right panel and then use the table of contents to switch to another slide of the same type the contents of the page are not updated properly this is an issue we had with dynamic panels in the main storylines app too has something to do with vue not re rendering everything steps to reproduce click on the oil sands extraction slide and switch to the right panel click on the in situ extraction slide notice that the image displayed is still for oil sands extraction demo | 0 |
214,874 | 24,121,154,245 | IssuesEvent | 2022-09-20 18:50:25 | Azure/AKS | https://api.github.com/repos/Azure/AKS | closed | Add Support for a KMS provider for Encrypting Secrets | security feature-request resolution/shipped | According to https://github.com/Azure/kubernetes-kms:
> We have added this feature to aks-engine so that you do not have to worry about any of the manual steps to set this up.
Does this mean Azure AKS encrypts secrets by default? I don't see a Key Vault being setup in the AKS resource group, so where are the keys stored? The above link also says:
> Currently KMS plugin for Kubernetes does not support KMS key rotation scenarios.
So keys are never rotated and hard coded from the start? | True | Add Support for a KMS provider for Encrypting Secrets - According to https://github.com/Azure/kubernetes-kms:
> We have added this feature to aks-engine so that you do not have to worry about any of the manual steps to set this up.
Does this mean Azure AKS encrypts secrets by default? I don't see a Key Vault being setup in the AKS resource group, so where are the keys stored? The above link also says:
> Currently KMS plugin for Kubernetes does not support KMS key rotation scenarios.
So keys are never rotated and hard coded from the start? | non_reli | add support for a kms provider for encrypting secrets according to we have added this feature to aks engine so that you do not have to worry about any of the manual steps to set this up does this mean azure aks encrypts secrets by default i don t see a key vault being setup in the aks resource group so where are the keys stored the above link also says currently kms plugin for kubernetes does not support kms key rotation scenarios so keys are never rotated and hard coded from the start | 0 |
137,202 | 20,101,659,101 | IssuesEvent | 2022-02-07 05:25:24 | Facepunch/sbox-issues | https://api.github.com/repos/Facepunch/sbox-issues | opened | Add WithSize, WithFormat, etc to TextureCubeBuilder | api design | **What it is**
Currently, TextureCubeBuilder only has the methods WithUAVBinding and Finish. Texture.CreateCube doesn't take any arguments either.
As far as I can tell, there's no way to specify the width/height of the cube texture, or the format of the data passed in.
As a result, it is currently impossible to create cube textures (spits out an error about the texture having 0x0 size).
**What it should be**
TextureCubeBuilder should have the relevant `WithSize(int width, int height)`, `WithFormat(ImageFormat format)`, `WithMips(int mips)`, etc methods, similar to Texture2DBuilder.
Ideally this functionality would also be added to TextureArrayBuilder. | 1.0 | Add WithSize, WithFormat, etc to TextureCubeBuilder - **What it is**
Currently, TextureCubeBuilder only has the methods WithUAVBinding and Finish. Texture.CreateCube doesn't take any arguments either.
As far as I can tell, there's no way to specify the width/height of the cube texture, or the format of the data passed in.
As a result, it is currently impossible to create cube textures (spits out an error about the texture having 0x0 size).
**What it should be**
TextureCubeBuilder should have the relevant `WithSize(int width, int height)`, `WithFormat(ImageFormat format)`, `WithMips(int mips)`, etc methods, similar to Texture2DBuilder.
Ideally this functionality would also be added to TextureArrayBuilder. | non_reli | add withsize withformat etc to texturecubebuilder what it is currently texturecubebuilder only has the methods withuavbinding and finish texture createcube doesn t take any arguments either as far as i can tell there s no way to specify the width height of the cube texture or the format of the data passed in as a result it is currently impossible to create cube textures spits out an error about the texture having size what it should be texturecubebuilder should have the relevant withsize int width int height withformat imageformat format withmips int mips etc methods similar to ideally this functionality would also be added to texturearraybuilder | 0 |
841 | 10,789,348,390 | IssuesEvent | 2019-11-05 11:43:06 | sohaibaslam/learning_site | https://api.github.com/repos/sohaibaslam/learning_site | opened | Broken Crawler: puma | crawler broken/unreliable | name: puma-at
name: puma-be
name: puma-ch
name: puma-cz
name: puma-de
name: puma-dk
name: puma-es
name: puma-fi
name: puma-fr
name: puma-gr
name: puma-hu
name: puma-ie
name: puma-it
name: puma-lu
name: puma-nl
name: puma-pl
name: puma-pt
name: puma-se
mixin_count: 18 | True | Broken Crawler: puma - name: puma-at
name: puma-be
name: puma-ch
name: puma-cz
name: puma-de
name: puma-dk
name: puma-es
name: puma-fi
name: puma-fr
name: puma-gr
name: puma-hu
name: puma-ie
name: puma-it
name: puma-lu
name: puma-nl
name: puma-pl
name: puma-pt
name: puma-se
mixin_count: 18 | reli | broken crawler puma name puma at name puma be name puma ch name puma cz name puma de name puma dk name puma es name puma fi name puma fr name puma gr name puma hu name puma ie name puma it name puma lu name puma nl name puma pl name puma pt name puma se mixin count | 1 |
165,437 | 14,002,060,981 | IssuesEvent | 2020-10-28 14:22:29 | AY2021S1-CS2103T-T09-1/tp | https://api.github.com/repos/AY2021S1-CS2103T-T09-1/tp | closed | UG: Add Tags command + update add, edit, find commands | documentation priority.High type.Task | - Add Tags command
- Update add command
- Update edit command
- Update find command | 1.0 | UG: Add Tags command + update add, edit, find commands - - Add Tags command
- Update add command
- Update edit command
- Update find command | non_reli | ug add tags command update add edit find commands add tags command update add command update edit command update find command | 0 |
259,874 | 8,200,923,745 | IssuesEvent | 2018-09-01 11:14:52 | TuranicTeam/Altay | https://api.github.com/repos/TuranicTeam/Altay | closed | Banner's | Category: Core Priority: Low Type: Cosmetic | #### Issue description
It would be necessary to operate banner crafts!
[

](url)
#### Steps to reproduce the issue
<!--- Help us find the problem by adding steps to reproduce the issue -->
-
#### OS and versions
<!---
Valid version must contain build number or git hash, "latest" is invalid.
If the version is invalid, the issue will be CLOSED
Game versions include Bedrock Engine, Windows 10 Edition version
-->
* Altay: 3.0.7+dev
* PHP: 7.2.8
* Server OS: Windows 10
* Game version: 1.5.2 Windows 10 Edition
#### Crashdump, backtrace, memory dumps, plugins list or other files
<!--- Paste in the below block -->
```
-
```
| 1.0 | Banner's - #### Issue description
It would be necessary to operate banner crafts!
[

](url)
#### Steps to reproduce the issue
<!--- Help us find the problem by adding steps to reproduce the issue -->
-
#### OS and versions
<!---
Valid version must contain build number or git hash, "latest" is invalid.
If the version is invalid, the issue will be CLOSED
Game versions include Bedrock Engine, Windows 10 Edition version
-->
* Altay: 3.0.7+dev
* PHP: 7.2.8
* Server OS: Windows 10
* Game version: 1.5.2 Windows 10 Edition
#### Crashdump, backtrace, memory dumps, plugins list or other files
<!--- Paste in the below block -->
```
-
```
| non_reli | banner s issue description it would be necessary to operate banner crafts url steps to reproduce the issue os and versions valid version must contain build number or git hash latest is invalid if the version is invalid the issue will be closed game versions include bedrock engine windows edition version altay dev php server os windows game version windows edition crashdump backtrace memory dumps plugins list or other files | 0 |
165,756 | 12,879,870,884 | IssuesEvent | 2020-07-12 01:26:52 | osquery/osquery | https://api.github.com/repos/osquery/osquery | closed | Create tests for the table `sharing_preferences` | good-first-issue macOS test | ## Create tests for the table `sharing_preferences`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/darwin/sharing_preferences.table`
Source files:
- `osquery/tables/system/darwin/sharing_preferences.cpp`
Table generating function: `genSharingPreferences()`
| 1.0 | Create tests for the table `sharing_preferences` - ## Create tests for the table `sharing_preferences`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/darwin/sharing_preferences.table`
Source files:
- `osquery/tables/system/darwin/sharing_preferences.cpp`
Table generating function: `genSharingPreferences()`
| non_reli | create tests for the table sharing preferences create tests for the table sharing preferences create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs darwin sharing preferences table source files osquery tables system darwin sharing preferences cpp table generating function gensharingpreferences | 0 |
154 | 4,628,109,909 | IssuesEvent | 2016-09-28 02:20:03 | UofSSpaceDesignTeam/rover-hardware | https://api.github.com/repos/UofSSpaceDesignTeam/rover-hardware | closed | Replace bad battery cell | Mandatory Power Reliability | One of the battery cells started expanding. It has been removed, and will be taken away to be (hopefully) disposed of properly. A new one will need to be installed to replace it. A new case will need to be designed to hold the cells under the right amount of pressure (job for the mechanical team?). | True | Replace bad battery cell - One of the battery cells started expanding. It has been removed, and will be taken away to be (hopefully) disposed of properly. A new one will need to be installed to replace it. A new case will need to be designed to hold the cells under the right amount of pressure (job for the mechanical team?). | reli | replace bad battery cell one of the battery cells started expanding it has been removed and will be taken away to be hopefully disposed of properly a new one will need to be installed to replace it a new case will need to be designed to hold the cells under the right amount of pressure job for the mechanical team | 1 |
471 | 7,734,135,141 | IssuesEvent | 2018-05-26 20:27:07 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Cloud provider api issues can prevent kubelet from reporting node status | area/reliability kind/bug lifecycle/rotten sig/node | There have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes. If the cloud provider api is unresponsive, it could prevent the node from reporting node/pod status to the apiserver. If this happens for too long, the node will be marked NotReady even though the kubelet is actually running without any issues.
I don't think we want to prevent a node or pod update because the cloud provider api has become unresponsive. We should try to ensure a node only gets marked NotReady if it is truly incapable of operating.
xref: #43516 #46602 #46340
@dchen1107 @kubernetes/sig-node-bugs | True | Cloud provider api issues can prevent kubelet from reporting node status - There have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes. If the cloud provider api is unresponsive, it could prevent the node from reporting node/pod status to the apiserver. If this happens for too long, the node will be marked NotReady even though the kubelet is actually running without any issues.
I don't think we want to prevent a node or pod update because the cloud provider api has become unresponsive. We should try to ensure a node only gets marked NotReady if it is truly incapable of operating.
xref: #43516 #46602 #46340
@dchen1107 @kubernetes/sig-node-bugs | reli | cloud provider api issues can prevent kubelet from reporting node status there have been a number of reported instances of cloud provider services becoming unresponsive and this causing adverse affects in kubernetes if the cloud provider api is unresponsive it could prevent the node from reporting node pod status to the apiserver if this happens for too long the node will be marked notready even though the kubelet is actually running without any issues i don t think we want to prevent a node or pod update because the cloud provider api has become unresponsive we should try to ensure a node only gets marked notready if it is truly incapable of operating xref kubernetes sig node bugs | 1 |
2,080 | 23,229,350,057 | IssuesEvent | 2022-08-03 05:50:57 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Infinite recursion during resource lookup within System.Private.CoreLib | area-Meta tenet-reliability untriaged needs-author-action |
### Description
```
CoreCLR Version: 6.0.622.26707
.NET Version: 6.0.6
Description: The application requested process termination through System.Environment.FailFast(string message).
Message: Infinite recursion during resource lookup within System.Private.CoreLib. This may be a bug in System.Private.CoreLib, or potentially in certain extensibility points such as assembly resolve events or CultureInfo names. Resource name: Arg_NullReferenceException
Stack:
at System.Environment.FailFast(System.String)
at System.SR.InternalGetResourceString(System.String)
at System.SR.GetResourceString(System.String)
at System.NullReferenceException..ctor()
at System.Globalization.CultureInfo.get_CurrentUICulture()
at System.Resources.ResourceManager.GetString(System.String, System.Globalization.CultureInfo)
at System.SR.InternalGetResourceString(System.String)
at System.SR.GetResourceString(System.String)
at System.NullReferenceException..ctor()
at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext(System.Threading.Thread)
at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext()
at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.InvokeContinuation(System.Action`1<System.Object>, System.Object, Boolean, Boolean)
at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.OnCompleted(System.Net.Sockets.SocketAsyncEventArgs)
at System.Threading._IOCompletionCallback.PerformIOCompletionCallback(UInt32, UInt32, System.Threading.NativeOverlapped*)
```
### Configuration
<!--
* Which version of .NET is the code running on?
* What OS and version, and for Linux, what distro?
* What is the architecture (x64, x86, ARM, ARM64)?
* Do you know whether it is specific to that configuration?
* If you're using Blazor, which web browser(s) do you see this issue in?
-->
### Regression?
<!--
* Did this work in a previous build or release of .NET, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK.
-->
### Other information
<!--
* Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches).
* If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of.
* Do you know of any workarounds?
-->
| True | Infinite recursion during resource lookup within System.Private.CoreLib -
### Description
```
CoreCLR Version: 6.0.622.26707
.NET Version: 6.0.6
Description: The application requested process termination through System.Environment.FailFast(string message).
Message: Infinite recursion during resource lookup within System.Private.CoreLib. This may be a bug in System.Private.CoreLib, or potentially in certain extensibility points such as assembly resolve events or CultureInfo names. Resource name: Arg_NullReferenceException
Stack:
at System.Environment.FailFast(System.String)
at System.SR.InternalGetResourceString(System.String)
at System.SR.GetResourceString(System.String)
at System.NullReferenceException..ctor()
at System.Globalization.CultureInfo.get_CurrentUICulture()
at System.Resources.ResourceManager.GetString(System.String, System.Globalization.CultureInfo)
at System.SR.InternalGetResourceString(System.String)
at System.SR.GetResourceString(System.String)
at System.NullReferenceException..ctor()
at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext(System.Threading.Thread)
at System.Runtime.CompilerServices.AsyncTaskMethodBuilder`1+AsyncStateMachineBox`1[[System.Int32, System.Private.CoreLib, Version=6.0.0.0, Culture=neutral, PublicKeyToken=7cec85d7bea7798e],[System.Net.Security.SslStream+<EnsureFullTlsFrameAsync>d__186`1[[System.Net.Security.AsyncReadWriteAdapter, System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]], System.Net.Security, Version=6.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a]].MoveNext()
at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.InvokeContinuation(System.Action`1<System.Object>, System.Object, Boolean, Boolean)
at System.Net.Sockets.Socket+AwaitableSocketAsyncEventArgs.OnCompleted(System.Net.Sockets.SocketAsyncEventArgs)
at System.Threading._IOCompletionCallback.PerformIOCompletionCallback(UInt32, UInt32, System.Threading.NativeOverlapped*)
```
### Configuration
<!--
* Which version of .NET is the code running on?
* What OS and version, and for Linux, what distro?
* What is the architecture (x64, x86, ARM, ARM64)?
* Do you know whether it is specific to that configuration?
* If you're using Blazor, which web browser(s) do you see this issue in?
-->
### Regression?
<!--
* Did this work in a previous build or release of .NET, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK.
-->
### Other information
<!--
* Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches).
* If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of.
* Do you know of any workarounds?
-->
| reli | infinite recursion during resource lookup within system private corelib description coreclr version net version description the application requested process termination through system environment failfast string message message infinite recursion during resource lookup within system private corelib this may be a bug in system private corelib or potentially in certain extensibility points such as assembly resolve events or cultureinfo names resource name arg nullreferenceexception stack at system environment failfast system string at system sr internalgetresourcestring system string at system sr getresourcestring system string at system nullreferenceexception ctor at system globalization cultureinfo get currentuiculture at system resources resourcemanager getstring system string system globalization cultureinfo at system sr internalgetresourcestring system string at system sr getresourcestring system string at system nullreferenceexception ctor at system threading executioncontext runinternal system threading executioncontext system threading contextcallback system object at system runtime compilerservices asynctaskmethodbuilder asyncstatemachinebox system net security version culture neutral publickeytoken movenext system threading thread at system runtime compilerservices asynctaskmethodbuilder asyncstatemachinebox system net security version culture neutral publickeytoken movenext at system net sockets socket awaitablesocketasynceventargs invokecontinuation system action system object boolean boolean at system net sockets socket awaitablesocketasynceventargs oncompleted system net sockets socketasynceventargs at system threading iocompletioncallback performiocompletioncallback system threading nativeoverlapped configuration which version of net is the code running on what os and version and for linux what distro what is the architecture arm do you know whether it is specific to that configuration if you re using blazor which web browser s do you see this issue in regression did this work in a previous build or release of net or from net framework if you can try a previous release or build to find out that can help us narrow down the problem if you don t know that s ok other information please include any relevant stack traces or error messages if possible please include text as text rather than images so it shows up in searches if you have an idea where the problem might lie let us know that here please include any pointers to code relevant changes or related issues you know of do you know of any workarounds | 1 |
273,459 | 8,530,913,133 | IssuesEvent | 2018-11-04 05:16:26 | CS2113-AY1819S1-W13-2/main | https://api.github.com/repos/CS2113-AY1819S1-W13-2/main | closed | Wrong prefix format used should reflect invalid command format message instead. | priority.high severity.High type.bug | **Describe the bug**
Wrong prefix format used should reflect invalid command format message instead.
**To Reproduce**
Steps to reproduce the behavior:
put additional prefix at the end of the full add command as showed in picture.
**Expected behavior**
A clear and concise description of what you expected to happen.
An invalid command format error message should be prompt instead of an internal error occured. This might confuse end user.
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other context about the problem here.
<hr>
**Reported by:** @ryanchen2018
**Severity:** Not Specified
<sub>[original: nusCS2113-AY1819S1/pe-1#154]</sub> | 1.0 | Wrong prefix format used should reflect invalid command format message instead. - **Describe the bug**
Wrong prefix format used should reflect invalid command format message instead.
**To Reproduce**
Steps to reproduce the behavior:
put additional prefix at the end of the full add command as showed in picture.
**Expected behavior**
A clear and concise description of what you expected to happen.
An invalid command format error message should be prompt instead of an internal error occured. This might confuse end user.
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other context about the problem here.
<hr>
**Reported by:** @ryanchen2018
**Severity:** Not Specified
<sub>[original: nusCS2113-AY1819S1/pe-1#154]</sub> | non_reli | wrong prefix format used should reflect invalid command format message instead describe the bug wrong prefix format used should reflect invalid command format message instead to reproduce steps to reproduce the behavior put additional prefix at the end of the full add command as showed in picture expected behavior a clear and concise description of what you expected to happen an invalid command format error message should be prompt instead of an internal error occured this might confuse end user screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reported by severity not specified | 0 |
281 | 6,001,208,880 | IssuesEvent | 2017-06-05 08:28:24 | datacite/datacite | https://api.github.com/repos/datacite/datacite | opened | Incident report have service outage | data center member reliability | As a data center manager (or member), I want a complete report after every incident so I can explain what happened to my users. | True | Incident report have service outage - As a data center manager (or member), I want a complete report after every incident so I can explain what happened to my users. | reli | incident report have service outage as a data center manager or member i want a complete report after every incident so i can explain what happened to my users | 1 |
792,040 | 27,943,705,604 | IssuesEvent | 2023-03-24 00:03:35 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | [MS] Create monitoring dashboard for MSPileup microservice | New Feature High Priority Stakeholders MSPileup QPrio: High | **Impact of the new feature**
New WMCore microservice
**Is your feature request related to a problem? Please describe.**
Creation of a new microservice to make pileup management more efficient and easier.
Subtask of https://github.com/dmwm/WMCore/issues/9779
**Describe the solution you'd like**
This is a joint ticket between the WMCore and the CMS Monitoring team, such that we can define and provide a monitoring dashboard for the MSPileup. So, relevant JIRA ticket might have to be created to get the usual process monitoring, as we have for the other WMCore microservices.
**Describe alternatives you've considered**
We can eventually consider service specific metrics as well, but I'd call that to be an future enhancement and we can properly define that in the future.
**Additional context**
None
| 1.0 | [MS] Create monitoring dashboard for MSPileup microservice - **Impact of the new feature**
New WMCore microservice
**Is your feature request related to a problem? Please describe.**
Creation of a new microservice to make pileup management more efficient and easier.
Subtask of https://github.com/dmwm/WMCore/issues/9779
**Describe the solution you'd like**
This is a joint ticket between the WMCore and the CMS Monitoring team, such that we can define and provide a monitoring dashboard for the MSPileup. So, relevant JIRA ticket might have to be created to get the usual process monitoring, as we have for the other WMCore microservices.
**Describe alternatives you've considered**
We can eventually consider service specific metrics as well, but I'd call that to be an future enhancement and we can properly define that in the future.
**Additional context**
None
| non_reli | create monitoring dashboard for mspileup microservice impact of the new feature new wmcore microservice is your feature request related to a problem please describe creation of a new microservice to make pileup management more efficient and easier subtask of describe the solution you d like this is a joint ticket between the wmcore and the cms monitoring team such that we can define and provide a monitoring dashboard for the mspileup so relevant jira ticket might have to be created to get the usual process monitoring as we have for the other wmcore microservices describe alternatives you ve considered we can eventually consider service specific metrics as well but i d call that to be an future enhancement and we can properly define that in the future additional context none | 0 |
336,273 | 24,491,492,483 | IssuesEvent | 2022-10-10 02:46:55 | vamsitadikonda/chat-defender-bot | https://api.github.com/repos/vamsitadikonda/chat-defender-bot | closed | Doc generation issues | documentation | Documentation generation is not syncing with GitHub pages. Please update the hook | 1.0 | Doc generation issues - Documentation generation is not syncing with GitHub pages. Please update the hook | non_reli | doc generation issues documentation generation is not syncing with github pages please update the hook | 0 |
2,840 | 28,101,638,438 | IssuesEvent | 2023-03-30 20:00:47 | supaglue-labs/supaglue | https://api.github.com/repos/supaglue-labs/supaglue | closed | Make sync history log ids idempotent | reliability | Right now, when we log a new SyncHistory event, we generate an id on the fly (or rather, prisma/DB does). This means that on retry, we create a new set of history log ids. Instead of doing this, we should not be writing another set of SyncHistory events.
We can do this by using deterministic UUIDs. We can do this by:
1. using temporal's deterministic UUID generator
2. Generate a "fake UUID" like "sync-id:iteration-number`
3. Generate UUIDv5 with namespace using iteration number | True | Make sync history log ids idempotent - Right now, when we log a new SyncHistory event, we generate an id on the fly (or rather, prisma/DB does). This means that on retry, we create a new set of history log ids. Instead of doing this, we should not be writing another set of SyncHistory events.
We can do this by using deterministic UUIDs. We can do this by:
1. using temporal's deterministic UUID generator
2. Generate a "fake UUID" like "sync-id:iteration-number`
3. Generate UUIDv5 with namespace using iteration number | reli | make sync history log ids idempotent right now when we log a new synchistory event we generate an id on the fly or rather prisma db does this means that on retry we create a new set of history log ids instead of doing this we should not be writing another set of synchistory events we can do this by using deterministic uuids we can do this by using temporal s deterministic uuid generator generate a fake uuid like sync id iteration number generate with namespace using iteration number | 1 |
6,181 | 13,886,002,489 | IssuesEvent | 2020-10-18 22:34:42 | bSchnepp/Feral | https://api.github.com/repos/bSchnepp/Feral | closed | [REGRESSION] - RS232 driver is no longer functioning | architecture feature high-priority | Since the CMake migration (https://github.com/bSchnepp/Feral/commit/79a9c3a2631900b422757ce159204135c7c2842b), it appears that the serial driver no longer functions, and COM1 is empty when the kernel finished booting. | 1.0 | [REGRESSION] - RS232 driver is no longer functioning - Since the CMake migration (https://github.com/bSchnepp/Feral/commit/79a9c3a2631900b422757ce159204135c7c2842b), it appears that the serial driver no longer functions, and COM1 is empty when the kernel finished booting. | non_reli | driver is no longer functioning since the cmake migration it appears that the serial driver no longer functions and is empty when the kernel finished booting | 0 |
448 | 7,573,968,079 | IssuesEvent | 2018-04-23 19:27:01 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Reduce allocation requirements in WriteObjectWorker | 4 - In Review Area-Compilers Bug Investigation Required Tenet-Reliability | The dictionaries used by `WriteObjectWorker` can become large, and occasionally fail when the IDE is under memory pressure. We should identify sets of objects which cannot be multiply-referenced within the object getting serialized, and avoid adding these objects to the map.
https://github.com/dotnet/roslyn/blob/878ffad23b8b06cb229c9ab31eada7634a473508/src/Compilers/Core/Portable/Serialization/ObjectWriter.cs#L805
:link: Watson (internal issue for this): https://devdiv.visualstudio.com/DevDiv/_workitems/edit/485537 | True | Reduce allocation requirements in WriteObjectWorker - The dictionaries used by `WriteObjectWorker` can become large, and occasionally fail when the IDE is under memory pressure. We should identify sets of objects which cannot be multiply-referenced within the object getting serialized, and avoid adding these objects to the map.
https://github.com/dotnet/roslyn/blob/878ffad23b8b06cb229c9ab31eada7634a473508/src/Compilers/Core/Portable/Serialization/ObjectWriter.cs#L805
:link: Watson (internal issue for this): https://devdiv.visualstudio.com/DevDiv/_workitems/edit/485537 | reli | reduce allocation requirements in writeobjectworker the dictionaries used by writeobjectworker can become large and occasionally fail when the ide is under memory pressure we should identify sets of objects which cannot be multiply referenced within the object getting serialized and avoid adding these objects to the map link watson internal issue for this | 1 |
622,429 | 19,635,153,302 | IssuesEvent | 2022-01-08 05:57:37 | Treetrain1/Survival-Remastered | https://api.github.com/repos/Treetrain1/Survival-Remastered | closed | On exiting the end portal, you will teleport into the void and die | bug high priority | When exiting the end through the portal using beta 5, you will probably fall through the world and die | 1.0 | On exiting the end portal, you will teleport into the void and die - When exiting the end through the portal using beta 5, you will probably fall through the world and die | non_reli | on exiting the end portal you will teleport into the void and die when exiting the end through the portal using beta you will probably fall through the world and die | 0 |
708 | 9,980,573,596 | IssuesEvent | 2019-07-10 04:11:33 | dotnet/project-system | https://api.github.com/repos/dotnet/project-system | closed | Renamed an item from an import crashes VS | Bug Feature-Project-File-Simplification Project-System-CPS Tenet-Reliability Urgency-Soon | Renamed an item from an import
```
Application: devenv.exe
Framework Version: v4.0.30319
Description: The process was terminated due to an unhandled exception.
Exception Info: System.NullReferenceException
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<IsCompilationCaseSensitiveAsync>d__15.MoveNext()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task)
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<RenameAsync>d__13.MoveNext()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.ValidateEnd(System.Threading.Tasks.Task)
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<OnWorkspaceChangedAsync>d__12.MoveNext()
at System.Runtime.CompilerServices.AsyncMethodBuilderCore+<>c.<ThrowAsync>b__6_0(System.Object)
at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32)
at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate)
at System.Windows.Threading.DispatcherOperation.InvokeImpl()
at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(System.Object)
at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(System.Object)
at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean)
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean)
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at MS.Internal.CulturePreservingExecutionContext.Run(MS.Internal.CulturePreservingExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Windows.Threading.DispatcherOperation.Invoke()
at System.Windows.Threading.Dispatcher.ProcessQueue()
at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef)
at MS.Win32.HwndWrapper.WndProc(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef)
at MS.Win32.HwndSubclass.DispatcherCallbackOperation(System.Object)
at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32)
at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate)
at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(System.Windows.Threading.DispatcherPriority, System.TimeSpan, System.Delegate, System.Object, Int32)
at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr, Int32, IntPtr, IntPtr)
``` | True | Renamed an item from an import crashes VS - Renamed an item from an import
```
Application: devenv.exe
Framework Version: v4.0.30319
Description: The process was terminated due to an unhandled exception.
Exception Info: System.NullReferenceException
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<IsCompilationCaseSensitiveAsync>d__15.MoveNext()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task)
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<RenameAsync>d__13.MoveNext()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(System.Threading.Tasks.Task)
at System.Runtime.CompilerServices.TaskAwaiter.ValidateEnd(System.Threading.Tasks.Task)
at Microsoft.VisualStudio.ProjectSystem.VS.Rename.Renamer+<OnWorkspaceChangedAsync>d__12.MoveNext()
at System.Runtime.CompilerServices.AsyncMethodBuilderCore+<>c.<ThrowAsync>b__6_0(System.Object)
at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32)
at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate)
at System.Windows.Threading.DispatcherOperation.InvokeImpl()
at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(System.Object)
at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(System.Object)
at System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean)
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object, Boolean)
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at MS.Internal.CulturePreservingExecutionContext.Run(MS.Internal.CulturePreservingExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Windows.Threading.DispatcherOperation.Invoke()
at System.Windows.Threading.Dispatcher.ProcessQueue()
at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef)
at MS.Win32.HwndWrapper.WndProc(IntPtr, Int32, IntPtr, IntPtr, Boolean ByRef)
at MS.Win32.HwndSubclass.DispatcherCallbackOperation(System.Object)
at System.Windows.Threading.ExceptionWrapper.InternalRealCall(System.Delegate, System.Object, Int32)
at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(System.Object, System.Delegate, System.Object, Int32, System.Delegate)
at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(System.Windows.Threading.DispatcherPriority, System.TimeSpan, System.Delegate, System.Object, Int32)
at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr, Int32, IntPtr, IntPtr)
``` | reli | renamed an item from an import crashes vs renamed an item from an import application devenv exe framework version description the process was terminated due to an unhandled exception exception info system nullreferenceexception at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices taskawaiter throwfornonsuccess system threading tasks task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification system threading tasks task at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices taskawaiter throwfornonsuccess system threading tasks task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification system threading tasks task at system runtime compilerservices taskawaiter validateend system threading tasks task at microsoft visualstudio projectsystem vs rename renamer d movenext at system runtime compilerservices asyncmethodbuildercore c b system object at system windows threading exceptionwrapper internalrealcall system delegate system object at system windows threading exceptionwrapper trycatchwhen system object system delegate system object system delegate at system windows threading dispatcheroperation invokeimpl at system windows threading dispatcheroperation invokeinsecuritycontext system object at ms internal culturepreservingexecutioncontext callbackwrapper system object at system threading executioncontext runinternal system threading executioncontext system threading contextcallback system object boolean at system threading executioncontext run system threading executioncontext system threading contextcallback system object boolean at system threading executioncontext run system threading executioncontext system threading contextcallback system object at ms internal culturepreservingexecutioncontext run ms internal culturepreservingexecutioncontext system threading contextcallback system object at system windows threading dispatcheroperation invoke at system windows threading dispatcher processqueue at system windows threading dispatcher wndprochook intptr intptr intptr boolean byref at ms hwndwrapper wndproc intptr intptr intptr boolean byref at ms hwndsubclass dispatchercallbackoperation system object at system windows threading exceptionwrapper internalrealcall system delegate system object at system windows threading exceptionwrapper trycatchwhen system object system delegate system object system delegate at system windows threading dispatcher legacyinvokeimpl system windows threading dispatcherpriority system timespan system delegate system object at ms hwndsubclass subclasswndproc intptr intptr intptr | 1 |
990 | 12,013,684,129 | IssuesEvent | 2020-04-10 09:27:25 | microsoft/calculator | https://api.github.com/repos/microsoft/calculator | closed | AltGr + E causes unhandled exception 0x40080201 | Area: Reliability Bug Pri: 2 triage approved | Calculator crashes when using the [`AltGr` key](https://en.wikipedia.org/wiki/AltGr_key) in combination with the `E` key.
**Steps To Reproduce**
1. Open calculator in any mode (for example `Standard`).
2. Type nothing or anything you want.
3. Press at least once or hold the `AltGr` key on a physical or virtual keyboard.
4. Type nothing or anything you want, except the `E` key.
5. Press the `E` key.
6. Calculator terminates with unhandled exception code `0x40080201`.
Pseudo-regex for the key presses after opening calc: `[^AltGr]* AltGr+ [^E]* E`.
**Device and Application Information**
- OS Build: 10.0.16299.0
- Architecture: x64
- Application Version: 10.1706.2406.0
- Region: de-DE
- Dev Version Installed: False
**Additional context**
- Stacktrace:
```
KERNELBASE.dll!RaiseException+0x68
combase.dll!ObjectStublessClient32+0x5b443
combase.dll!RoOriginateError+0x51
vccorlib140_app.DLL!Platform::Details::EventSourceRemove+0x1239
vccorlib140_app.DLL!__abi_WinRTraiseInvalidCastException+0x17
vccorlib140_app.DLL!Platform::String::<Dispose>+0x1e8
vccorlib140_app.DLL!Platform::Details::ResolveWeakReference+0x98
Calculator.exe+0x3bd4e
Calculator.exe+0x42b61
Windows.UI.dll!Ordinal1689+0x6e21
Windows.UI.dll!Ordinal1689+0x489e
windows.ui.core.textinput.dll!Ordinal1501+0x31d0
windows.ui.core.textinput.dll!Ordinal1501+0x2a2b
TextInputFramework.dll!TextInputHostSiteCreate+0x5dc1
TextInputFramework.dll!TextInputHostSiteCreate+0x82b0
TextInputFramework.dll!TextInputHostSiteCreate+0x5681
msctf.dll+0x3131
TextInputFramework.dll+0x18dc
TextInputFramework.dll!TextInputHostCreateEx+0x1499f
TextInputFramework.dll!TextInputHostCreateEx+0x14847
CoreMessaging.dll!CoreUICreateEx+0x2a8e
CoreMessaging.dll!CoreUICallReceive+0x499
CoreMessaging.dll!CoreUICreateEx+0x12eef
CoreMessaging.dll!CoreUICreateEx+0x1321a
CoreMessaging.dll!CoreUICreateEx+0x1077b
CoreMessaging.dll!CoreUICreateEx+0xbcc
CoreMessaging.dll!CoreUICallReceive+0xca99
CoreMessaging.dll!CoreUICallReceive+0xcc88
CoreMessaging.dll!CoreUICallReceive+0xd186
CoreMessaging.dll!CoreUICallReceive+0xd4d4
CoreMessaging.dll!CoreUICreateEx+0x10a6a
CoreMessaging.dll!CoreUICreateEx+0xffdc
CoreMessaging.dll!CoreUICreateEx+0x7fbd
CoreMessaging.dll!CoreUICreateEx+0x6527
CoreMessaging.dll!CoreUICreateEx+0x2d39
CoreMessaging.dll!CoreUICreateEx+0x2b19
user32.dll!CallWindowProcW+0x4dd
user32.dll!CallWindowProcW+0x1cc
user32.dll!GetTopWindow+0x153
ntdll.dll!KiUserCallbackDispatcher+0x24
win32u.dll!NtUserPeekMessage+0x14
user32.dll!PeekMessageW+0x88
Windows.UI.dll!Ordinal1500+0xd9f6
Windows.UI.dll!Ordinal1500+0xbbee
Windows.UI.dll!Ordinal1500+0xd89b
Windows.UI.Xaml.dll!DllGetActivationFactory+0x1856bc
Windows.UI.Xaml.dll!DllGetActivationFactory+0x185660
twinapi.appcore.dll!Ordinal7+0x306a
twinapi.appcore.dll!DllGetActivationFactory+0x146c
shcore.dll!Ordinal247+0x1b5
KERNEL32.DLL!BaseThreadInitThunk+0x14
ntdll.dll!RtlUserThreadStart+0x21
```
**Requested Assignment**
I'm just reporting this problem. I don't want to fix it. | True | AltGr + E causes unhandled exception 0x40080201 - Calculator crashes when using the [`AltGr` key](https://en.wikipedia.org/wiki/AltGr_key) in combination with the `E` key.
**Steps To Reproduce**
1. Open calculator in any mode (for example `Standard`).
2. Type nothing or anything you want.
3. Press at least once or hold the `AltGr` key on a physical or virtual keyboard.
4. Type nothing or anything you want, except the `E` key.
5. Press the `E` key.
6. Calculator terminates with unhandled exception code `0x40080201`.
Pseudo-regex for the key presses after opening calc: `[^AltGr]* AltGr+ [^E]* E`.
**Device and Application Information**
- OS Build: 10.0.16299.0
- Architecture: x64
- Application Version: 10.1706.2406.0
- Region: de-DE
- Dev Version Installed: False
**Additional context**
- Stacktrace:
```
KERNELBASE.dll!RaiseException+0x68
combase.dll!ObjectStublessClient32+0x5b443
combase.dll!RoOriginateError+0x51
vccorlib140_app.DLL!Platform::Details::EventSourceRemove+0x1239
vccorlib140_app.DLL!__abi_WinRTraiseInvalidCastException+0x17
vccorlib140_app.DLL!Platform::String::<Dispose>+0x1e8
vccorlib140_app.DLL!Platform::Details::ResolveWeakReference+0x98
Calculator.exe+0x3bd4e
Calculator.exe+0x42b61
Windows.UI.dll!Ordinal1689+0x6e21
Windows.UI.dll!Ordinal1689+0x489e
windows.ui.core.textinput.dll!Ordinal1501+0x31d0
windows.ui.core.textinput.dll!Ordinal1501+0x2a2b
TextInputFramework.dll!TextInputHostSiteCreate+0x5dc1
TextInputFramework.dll!TextInputHostSiteCreate+0x82b0
TextInputFramework.dll!TextInputHostSiteCreate+0x5681
msctf.dll+0x3131
TextInputFramework.dll+0x18dc
TextInputFramework.dll!TextInputHostCreateEx+0x1499f
TextInputFramework.dll!TextInputHostCreateEx+0x14847
CoreMessaging.dll!CoreUICreateEx+0x2a8e
CoreMessaging.dll!CoreUICallReceive+0x499
CoreMessaging.dll!CoreUICreateEx+0x12eef
CoreMessaging.dll!CoreUICreateEx+0x1321a
CoreMessaging.dll!CoreUICreateEx+0x1077b
CoreMessaging.dll!CoreUICreateEx+0xbcc
CoreMessaging.dll!CoreUICallReceive+0xca99
CoreMessaging.dll!CoreUICallReceive+0xcc88
CoreMessaging.dll!CoreUICallReceive+0xd186
CoreMessaging.dll!CoreUICallReceive+0xd4d4
CoreMessaging.dll!CoreUICreateEx+0x10a6a
CoreMessaging.dll!CoreUICreateEx+0xffdc
CoreMessaging.dll!CoreUICreateEx+0x7fbd
CoreMessaging.dll!CoreUICreateEx+0x6527
CoreMessaging.dll!CoreUICreateEx+0x2d39
CoreMessaging.dll!CoreUICreateEx+0x2b19
user32.dll!CallWindowProcW+0x4dd
user32.dll!CallWindowProcW+0x1cc
user32.dll!GetTopWindow+0x153
ntdll.dll!KiUserCallbackDispatcher+0x24
win32u.dll!NtUserPeekMessage+0x14
user32.dll!PeekMessageW+0x88
Windows.UI.dll!Ordinal1500+0xd9f6
Windows.UI.dll!Ordinal1500+0xbbee
Windows.UI.dll!Ordinal1500+0xd89b
Windows.UI.Xaml.dll!DllGetActivationFactory+0x1856bc
Windows.UI.Xaml.dll!DllGetActivationFactory+0x185660
twinapi.appcore.dll!Ordinal7+0x306a
twinapi.appcore.dll!DllGetActivationFactory+0x146c
shcore.dll!Ordinal247+0x1b5
KERNEL32.DLL!BaseThreadInitThunk+0x14
ntdll.dll!RtlUserThreadStart+0x21
```
**Requested Assignment**
I'm just reporting this problem. I don't want to fix it. | reli | altgr e causes unhandled exception calculator crashes when using the in combination with the e key steps to reproduce open calculator in any mode for example standard type nothing or anything you want press at least once or hold the altgr key on a physical or virtual keyboard type nothing or anything you want except the e key press the e key calculator terminates with unhandled exception code pseudo regex for the key presses after opening calc altgr e device and application information os build architecture application version region de de dev version installed false additional context stacktrace kernelbase dll raiseexception combase dll combase dll rooriginateerror app dll platform details eventsourceremove app dll abi winrtraiseinvalidcastexception app dll platform string app dll platform details resolveweakreference calculator exe calculator exe windows ui dll windows ui dll windows ui core textinput dll windows ui core textinput dll textinputframework dll textinputhostsitecreate textinputframework dll textinputhostsitecreate textinputframework dll textinputhostsitecreate msctf dll textinputframework dll textinputframework dll textinputhostcreateex textinputframework dll textinputhostcreateex coremessaging dll coreuicreateex coremessaging dll coreuicallreceive coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicallreceive coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex coremessaging dll coreuicreateex dll callwindowprocw dll callwindowprocw dll gettopwindow ntdll dll kiusercallbackdispatcher dll ntuserpeekmessage dll peekmessagew windows ui dll windows ui dll windows ui dll windows ui xaml dll dllgetactivationfactory windows ui xaml dll dllgetactivationfactory twinapi appcore dll twinapi appcore dll dllgetactivationfactory shcore dll dll basethreadinitthunk ntdll dll rtluserthreadstart requested assignment i m just reporting this problem i don t want to fix it | 1 |
14,504 | 2,814,133,886 | IssuesEvent | 2015-05-18 18:19:17 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | [TEST-FAILURE] ClientMemberAttributeTest.testChangeMemberAttributes | Team: Client Type: Defect | ```
java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds , count left: 1
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:603)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:596)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:588)
at com.hazelcast.client.ClientMemberAttributeTest.testChangeMemberAttributes(ClientMemberAttributeTest.java:76)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client-new/405/testReport/junit/com.hazelcast.client/ClientMemberAttributeTest/testChangeMemberAttributes/ | 1.0 | [TEST-FAILURE] ClientMemberAttributeTest.testChangeMemberAttributes - ```
java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds , count left: 1
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:603)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:596)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:588)
at com.hazelcast.client.ClientMemberAttributeTest.testChangeMemberAttributes(ClientMemberAttributeTest.java:76)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client-new/405/testReport/junit/com.hazelcast.client/ClientMemberAttributeTest/testChangeMemberAttributes/ | non_reli | clientmemberattributetest testchangememberattributes java lang assertionerror countdownlatch failed to complete within seconds count left at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast client clientmemberattributetest testchangememberattributes clientmemberattributetest java | 0 |
236 | 5,530,185,596 | IssuesEvent | 2017-03-21 01:25:35 | Azure/azure-webjobs-sdk-script | https://api.github.com/repos/Azure/azure-webjobs-sdk-script | closed | Scale controller needs more complex heuristic to estimate memory needs | reliability scale controller | There is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site, and the last writer wins. This results in situations where if one of the site’s instances is ‘inactive’, or under low load, let’s say due to inability to acquire some kind of lease, or being assigned to a non-busy partition, then, that worker will observe low a working set value, while a busy instance will have a (much) higher working set value.
A better metric would be to keep the last ‘n’ readings (e.g. 5 or 10), and always use the highest. It’s still fairly simple, but should perform a lot better.
Discussed in mail thread "app memory usage and instance scaling" | True | Scale controller needs more complex heuristic to estimate memory needs - There is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site, and the last writer wins. This results in situations where if one of the site’s instances is ‘inactive’, or under low load, let’s say due to inability to acquire some kind of lease, or being assigned to a non-busy partition, then, that worker will observe low a working set value, while a busy instance will have a (much) higher working set value.
A better metric would be to keep the last ‘n’ readings (e.g. 5 or 10), and always use the highest. It’s still fairly simple, but should perform a lot better.
Discussed in mail thread "app memory usage and instance scaling" | reli | scale controller needs more complex heuristic to estimate memory needs there is an issue in our current implementation that each worker that is running the site independently reports the active working set that it has been observing for site and the last writer wins this results in situations where if one of the site’s instances is ‘inactive’ or under low load let’s say due to inability to acquire some kind of lease or being assigned to a non busy partition then that worker will observe low a working set value while a busy instance will have a much higher working set value a better metric would be to keep the last ‘n’ readings e g or and always use the highest it’s still fairly simple but should perform a lot better discussed in mail thread app memory usage and instance scaling | 1 |
187,857 | 14,433,311,848 | IssuesEvent | 2020-12-07 04:29:54 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | reopened | Windows LambdaLoadTest hang | test failure | ERROR: type should be string, got "https://ci.eclipse.org/openj9/job/Test_openjdk8_j9_special.system_x86-32_windows_Personal/33\r\nLambdaLoadTest_OpenJ9_NonLinux_special_24\r\nvariation: Mode107-OSRG\r\nJVM_OPTIONS: -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation \r\n\r\nNo diagnostic files generated.\r\n\r\n```\r\n10:22:03.554 - Completed 3.0%. Number of tests started=6\r\n10:22:23.882 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:22:43.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:23:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:03.413 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:24:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:03.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:23.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:25:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:26:43.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:23.460 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:27:43.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:28:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:03.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:29:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:30:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:31:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:32:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:03.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:33:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:34:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:23.429 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:35:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:23.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:36:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:37:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:38:43.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:39:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:23.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:40:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:03.476 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:41:43.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:42:43.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:43:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:03.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:44:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:03.445 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:45:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:03.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:23.398 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:46:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:23.508 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:47:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:48:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:49:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:23.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:50:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:51:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:52:43.492 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:23.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:53:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:54:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:55:43.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:56:43.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:57:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:03.414 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:58:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:03.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n10:59:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:00:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:23.461 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:01:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:02:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:03:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:04:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:23.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:05:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:06:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:07:43.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:23.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:08:43.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:09:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:03.430 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:10:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:03.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:11:43.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:23.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:12:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:23.477 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:13:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:03.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:14:43.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:23.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:15:43.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:03.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:23.399 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:16:43.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:03.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:23.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:17:43.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:03.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:23.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:18:43.400 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:03.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:23.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:19:43.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:03.462 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:23.478 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:20:43.493 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:03.400 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:23.415 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:21:43.431 - Completed 3.0%. Number of tests started=6 (+0)\r\n11:22:03.446 - Completed 3.0%. Number of tests started=6 (+0)\r\n```" | 1.0 | Windows LambdaLoadTest hang - https://ci.eclipse.org/openj9/job/Test_openjdk8_j9_special.system_x86-32_windows_Personal/33
LambdaLoadTest_OpenJ9_NonLinux_special_24
variation: Mode107-OSRG
JVM_OPTIONS: -Xgcpolicy:optthruput -Xdebug -Xrunjdwp:transport=dt_socket,address=8888,server=y,onthrow=no.pkg.foo,launch=echo -Xjit:enableOSR,enableOSROnGuardFailure,count=1,disableAsyncCompilation
No diagnostic files generated.
```
10:22:03.554 - Completed 3.0%. Number of tests started=6
10:22:23.882 - Completed 3.0%. Number of tests started=6 (+0)
10:22:43.460 - Completed 3.0%. Number of tests started=6 (+0)
10:23:03.476 - Completed 3.0%. Number of tests started=6 (+0)
10:23:23.492 - Completed 3.0%. Number of tests started=6 (+0)
10:23:43.398 - Completed 3.0%. Number of tests started=6 (+0)
10:24:03.413 - Completed 3.0%. Number of tests started=6 (+0)
10:24:23.429 - Completed 3.0%. Number of tests started=6 (+0)
10:24:43.445 - Completed 3.0%. Number of tests started=6 (+0)
10:25:03.460 - Completed 3.0%. Number of tests started=6 (+0)
10:25:23.476 - Completed 3.0%. Number of tests started=6 (+0)
10:25:43.492 - Completed 3.0%. Number of tests started=6 (+0)
10:26:03.398 - Completed 3.0%. Number of tests started=6 (+0)
10:26:23.414 - Completed 3.0%. Number of tests started=6 (+0)
10:26:43.429 - Completed 3.0%. Number of tests started=6 (+0)
10:27:03.445 - Completed 3.0%. Number of tests started=6 (+0)
10:27:23.460 - Completed 3.0%. Number of tests started=6 (+0)
10:27:43.476 - Completed 3.0%. Number of tests started=6 (+0)
10:28:03.492 - Completed 3.0%. Number of tests started=6 (+0)
10:28:23.398 - Completed 3.0%. Number of tests started=6 (+0)
10:28:43.414 - Completed 3.0%. Number of tests started=6 (+0)
10:29:03.429 - Completed 3.0%. Number of tests started=6 (+0)
10:29:23.445 - Completed 3.0%. Number of tests started=6 (+0)
10:29:43.461 - Completed 3.0%. Number of tests started=6 (+0)
10:30:03.476 - Completed 3.0%. Number of tests started=6 (+0)
10:30:23.492 - Completed 3.0%. Number of tests started=6 (+0)
10:30:43.398 - Completed 3.0%. Number of tests started=6 (+0)
10:31:03.414 - Completed 3.0%. Number of tests started=6 (+0)
10:31:23.429 - Completed 3.0%. Number of tests started=6 (+0)
10:31:43.445 - Completed 3.0%. Number of tests started=6 (+0)
10:32:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:32:23.414 - Completed 3.0%. Number of tests started=6 (+0)
10:32:43.398 - Completed 3.0%. Number of tests started=6 (+0)
10:33:03.429 - Completed 3.0%. Number of tests started=6 (+0)
10:33:23.445 - Completed 3.0%. Number of tests started=6 (+0)
10:33:43.461 - Completed 3.0%. Number of tests started=6 (+0)
10:34:03.476 - Completed 3.0%. Number of tests started=6 (+0)
10:34:23.492 - Completed 3.0%. Number of tests started=6 (+0)
10:34:43.398 - Completed 3.0%. Number of tests started=6 (+0)
10:35:03.414 - Completed 3.0%. Number of tests started=6 (+0)
10:35:23.429 - Completed 3.0%. Number of tests started=6 (+0)
10:35:43.445 - Completed 3.0%. Number of tests started=6 (+0)
10:36:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:36:23.476 - Completed 3.0%. Number of tests started=6 (+0)
10:36:43.492 - Completed 3.0%. Number of tests started=6 (+0)
10:37:03.398 - Completed 3.0%. Number of tests started=6 (+0)
10:37:23.414 - Completed 3.0%. Number of tests started=6 (+0)
10:37:43.430 - Completed 3.0%. Number of tests started=6 (+0)
10:38:03.445 - Completed 3.0%. Number of tests started=6 (+0)
10:38:23.461 - Completed 3.0%. Number of tests started=6 (+0)
10:38:43.476 - Completed 3.0%. Number of tests started=6 (+0)
10:39:03.492 - Completed 3.0%. Number of tests started=6 (+0)
10:39:23.398 - Completed 3.0%. Number of tests started=6 (+0)
10:39:43.414 - Completed 3.0%. Number of tests started=6 (+0)
10:40:03.430 - Completed 3.0%. Number of tests started=6 (+0)
10:40:23.445 - Completed 3.0%. Number of tests started=6 (+0)
10:40:43.461 - Completed 3.0%. Number of tests started=6 (+0)
10:41:03.476 - Completed 3.0%. Number of tests started=6 (+0)
10:41:23.492 - Completed 3.0%. Number of tests started=6 (+0)
10:41:43.398 - Completed 3.0%. Number of tests started=6 (+0)
10:42:03.414 - Completed 3.0%. Number of tests started=6 (+0)
10:42:23.430 - Completed 3.0%. Number of tests started=6 (+0)
10:42:43.445 - Completed 3.0%. Number of tests started=6 (+0)
10:43:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:43:23.477 - Completed 3.0%. Number of tests started=6 (+0)
10:43:43.492 - Completed 3.0%. Number of tests started=6 (+0)
10:44:03.398 - Completed 3.0%. Number of tests started=6 (+0)
10:44:23.414 - Completed 3.0%. Number of tests started=6 (+0)
10:44:43.430 - Completed 3.0%. Number of tests started=6 (+0)
10:45:03.445 - Completed 3.0%. Number of tests started=6 (+0)
10:45:23.461 - Completed 3.0%. Number of tests started=6 (+0)
10:45:43.477 - Completed 3.0%. Number of tests started=6 (+0)
10:46:03.492 - Completed 3.0%. Number of tests started=6 (+0)
10:46:23.398 - Completed 3.0%. Number of tests started=6 (+0)
10:46:43.414 - Completed 3.0%. Number of tests started=6 (+0)
10:47:03.430 - Completed 3.0%. Number of tests started=6 (+0)
10:47:23.508 - Completed 3.0%. Number of tests started=6 (+0)
10:47:43.430 - Completed 3.0%. Number of tests started=6 (+0)
10:48:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:48:23.399 - Completed 3.0%. Number of tests started=6 (+0)
10:48:43.414 - Completed 3.0%. Number of tests started=6 (+0)
10:49:03.430 - Completed 3.0%. Number of tests started=6 (+0)
10:49:23.446 - Completed 3.0%. Number of tests started=6 (+0)
10:49:43.461 - Completed 3.0%. Number of tests started=6 (+0)
10:50:03.477 - Completed 3.0%. Number of tests started=6 (+0)
10:50:23.492 - Completed 3.0%. Number of tests started=6 (+0)
10:50:43.399 - Completed 3.0%. Number of tests started=6 (+0)
10:51:03.414 - Completed 3.0%. Number of tests started=6 (+0)
10:51:23.430 - Completed 3.0%. Number of tests started=6 (+0)
10:51:43.446 - Completed 3.0%. Number of tests started=6 (+0)
10:52:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:52:23.477 - Completed 3.0%. Number of tests started=6 (+0)
10:52:43.492 - Completed 3.0%. Number of tests started=6 (+0)
10:53:03.399 - Completed 3.0%. Number of tests started=6 (+0)
10:53:23.414 - Completed 3.0%. Number of tests started=6 (+0)
10:53:43.430 - Completed 3.0%. Number of tests started=6 (+0)
10:54:03.446 - Completed 3.0%. Number of tests started=6 (+0)
10:54:23.461 - Completed 3.0%. Number of tests started=6 (+0)
10:54:43.477 - Completed 3.0%. Number of tests started=6 (+0)
10:55:03.493 - Completed 3.0%. Number of tests started=6 (+0)
10:55:23.399 - Completed 3.0%. Number of tests started=6 (+0)
10:55:43.414 - Completed 3.0%. Number of tests started=6 (+0)
10:56:03.430 - Completed 3.0%. Number of tests started=6 (+0)
10:56:23.446 - Completed 3.0%. Number of tests started=6 (+0)
10:56:43.461 - Completed 3.0%. Number of tests started=6 (+0)
10:57:03.477 - Completed 3.0%. Number of tests started=6 (+0)
10:57:23.493 - Completed 3.0%. Number of tests started=6 (+0)
10:57:43.399 - Completed 3.0%. Number of tests started=6 (+0)
10:58:03.414 - Completed 3.0%. Number of tests started=6 (+0)
10:58:23.430 - Completed 3.0%. Number of tests started=6 (+0)
10:58:43.446 - Completed 3.0%. Number of tests started=6 (+0)
10:59:03.461 - Completed 3.0%. Number of tests started=6 (+0)
10:59:23.477 - Completed 3.0%. Number of tests started=6 (+0)
10:59:43.493 - Completed 3.0%. Number of tests started=6 (+0)
11:00:03.399 - Completed 3.0%. Number of tests started=6 (+0)
11:00:23.415 - Completed 3.0%. Number of tests started=6 (+0)
11:00:43.430 - Completed 3.0%. Number of tests started=6 (+0)
11:01:03.446 - Completed 3.0%. Number of tests started=6 (+0)
11:01:23.461 - Completed 3.0%. Number of tests started=6 (+0)
11:01:43.477 - Completed 3.0%. Number of tests started=6 (+0)
11:02:03.493 - Completed 3.0%. Number of tests started=6 (+0)
11:02:23.399 - Completed 3.0%. Number of tests started=6 (+0)
11:02:43.415 - Completed 3.0%. Number of tests started=6 (+0)
11:03:03.430 - Completed 3.0%. Number of tests started=6 (+0)
11:03:23.446 - Completed 3.0%. Number of tests started=6 (+0)
11:03:43.462 - Completed 3.0%. Number of tests started=6 (+0)
11:04:03.477 - Completed 3.0%. Number of tests started=6 (+0)
11:04:23.493 - Completed 3.0%. Number of tests started=6 (+0)
11:04:43.399 - Completed 3.0%. Number of tests started=6 (+0)
11:05:03.415 - Completed 3.0%. Number of tests started=6 (+0)
11:05:23.430 - Completed 3.0%. Number of tests started=6 (+0)
11:05:43.446 - Completed 3.0%. Number of tests started=6 (+0)
11:06:03.462 - Completed 3.0%. Number of tests started=6 (+0)
11:06:23.477 - Completed 3.0%. Number of tests started=6 (+0)
11:06:43.493 - Completed 3.0%. Number of tests started=6 (+0)
11:07:03.399 - Completed 3.0%. Number of tests started=6 (+0)
11:07:23.415 - Completed 3.0%. Number of tests started=6 (+0)
11:07:43.430 - Completed 3.0%. Number of tests started=6 (+0)
11:08:03.446 - Completed 3.0%. Number of tests started=6 (+0)
11:08:23.462 - Completed 3.0%. Number of tests started=6 (+0)
11:08:43.477 - Completed 3.0%. Number of tests started=6 (+0)
11:09:03.493 - Completed 3.0%. Number of tests started=6 (+0)
11:09:23.399 - Completed 3.0%. Number of tests started=6 (+0)
11:09:43.415 - Completed 3.0%. Number of tests started=6 (+0)
11:10:03.430 - Completed 3.0%. Number of tests started=6 (+0)
11:10:23.446 - Completed 3.0%. Number of tests started=6 (+0)
11:10:43.462 - Completed 3.0%. Number of tests started=6 (+0)
11:11:03.477 - Completed 3.0%. Number of tests started=6 (+0)
11:11:23.493 - Completed 3.0%. Number of tests started=6 (+0)
11:11:43.399 - Completed 3.0%. Number of tests started=6 (+0)
11:12:03.415 - Completed 3.0%. Number of tests started=6 (+0)
11:12:23.431 - Completed 3.0%. Number of tests started=6 (+0)
11:12:43.446 - Completed 3.0%. Number of tests started=6 (+0)
11:13:03.462 - Completed 3.0%. Number of tests started=6 (+0)
11:13:23.477 - Completed 3.0%. Number of tests started=6 (+0)
11:13:43.493 - Completed 3.0%. Number of tests started=6 (+0)
11:14:03.399 - Completed 3.0%. Number of tests started=6 (+0)
11:14:23.415 - Completed 3.0%. Number of tests started=6 (+0)
11:14:43.431 - Completed 3.0%. Number of tests started=6 (+0)
11:15:03.446 - Completed 3.0%. Number of tests started=6 (+0)
11:15:23.462 - Completed 3.0%. Number of tests started=6 (+0)
11:15:43.478 - Completed 3.0%. Number of tests started=6 (+0)
11:16:03.493 - Completed 3.0%. Number of tests started=6 (+0)
11:16:23.399 - Completed 3.0%. Number of tests started=6 (+0)
11:16:43.415 - Completed 3.0%. Number of tests started=6 (+0)
11:17:03.431 - Completed 3.0%. Number of tests started=6 (+0)
11:17:23.446 - Completed 3.0%. Number of tests started=6 (+0)
11:17:43.462 - Completed 3.0%. Number of tests started=6 (+0)
11:18:03.478 - Completed 3.0%. Number of tests started=6 (+0)
11:18:23.493 - Completed 3.0%. Number of tests started=6 (+0)
11:18:43.400 - Completed 3.0%. Number of tests started=6 (+0)
11:19:03.415 - Completed 3.0%. Number of tests started=6 (+0)
11:19:23.431 - Completed 3.0%. Number of tests started=6 (+0)
11:19:43.446 - Completed 3.0%. Number of tests started=6 (+0)
11:20:03.462 - Completed 3.0%. Number of tests started=6 (+0)
11:20:23.478 - Completed 3.0%. Number of tests started=6 (+0)
11:20:43.493 - Completed 3.0%. Number of tests started=6 (+0)
11:21:03.400 - Completed 3.0%. Number of tests started=6 (+0)
11:21:23.415 - Completed 3.0%. Number of tests started=6 (+0)
11:21:43.431 - Completed 3.0%. Number of tests started=6 (+0)
11:22:03.446 - Completed 3.0%. Number of tests started=6 (+0)
``` | non_reli | windows lambdaloadtest hang lambdaloadtest nonlinux special variation osrg jvm options xgcpolicy optthruput xdebug xrunjdwp transport dt socket address server y onthrow no pkg foo launch echo xjit enableosr enableosronguardfailure count disableasynccompilation no diagnostic files generated completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started completed number of tests started | 0 |
2,997 | 30,833,054,715 | IssuesEvent | 2023-08-02 04:29:16 | hackforla/ops | https://api.github.com/repos/hackforla/ops | opened | [SPIKE] Set up AWS CloudWatch billing alarms | size: 2pt role: Site Reliability Engineer feature: monitoring | ### Overview
Currently, we have no safe guards in place if an AWS service costs more than anticipated. AWS has measures in place with tools like Cloudwatch. Let's investigate how to set up a cloudwatch alarm and see if we can set up safeguards.
### Action Items
- [ ] Investigate HFLA's incubator AWS setup
- [ ] Determine a general implementation and potential costs
- [ ] Create implementation issue
- [ ] Create CloudWatch documentation issue
### Resources/Instructions
https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/monitor_estimated_charges_with_cloudwatch.html
| True | [SPIKE] Set up AWS CloudWatch billing alarms - ### Overview
Currently, we have no safe guards in place if an AWS service costs more than anticipated. AWS has measures in place with tools like Cloudwatch. Let's investigate how to set up a cloudwatch alarm and see if we can set up safeguards.
### Action Items
- [ ] Investigate HFLA's incubator AWS setup
- [ ] Determine a general implementation and potential costs
- [ ] Create implementation issue
- [ ] Create CloudWatch documentation issue
### Resources/Instructions
https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/monitor_estimated_charges_with_cloudwatch.html
| reli | set up aws cloudwatch billing alarms overview currently we have no safe guards in place if an aws service costs more than anticipated aws has measures in place with tools like cloudwatch let s investigate how to set up a cloudwatch alarm and see if we can set up safeguards action items investigate hfla s incubator aws setup determine a general implementation and potential costs create implementation issue create cloudwatch documentation issue resources instructions | 1 |
466 | 7,698,594,608 | IssuesEvent | 2018-05-19 00:41:51 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | WinHttpRequestState memory leak when cancelling requests | area-System.Net.Http bug os-windows tenet-reliability | (reported by ASP.NET team).
cc: @DavidObando
There is a memory leak of WinHttpRequestState (and related objects) when the outstanding http request is canceled prior to it completing with an HttpResponseMessage.
| True | WinHttpRequestState memory leak when cancelling requests - (reported by ASP.NET team).
cc: @DavidObando
There is a memory leak of WinHttpRequestState (and related objects) when the outstanding http request is canceled prior to it completing with an HttpResponseMessage.
| reli | winhttprequeststate memory leak when cancelling requests reported by asp net team cc davidobando there is a memory leak of winhttprequeststate and related objects when the outstanding http request is canceled prior to it completing with an httpresponsemessage | 1 |
2,851 | 28,236,391,837 | IssuesEvent | 2023-04-06 01:09:11 | hackforla/ops | https://api.github.com/repos/hackforla/ops | opened | Provision RDS database for People Depot dev environment | size: 2pt role: Data Engineer role: Site Reliability Engineer feature: deployment | ### Overview
As a People Depot developer, I'd like an developer environment database provisioned for People Depot developers to work against.
### Action Items
- [ ] Provision PostgresQL RDS
- [ ] Give dev credentials to People Depot
### Resources/Instructions
https://repost.aws/knowledge-center/ecs-fargate-task-database-connection
| True | Provision RDS database for People Depot dev environment - ### Overview
As a People Depot developer, I'd like an developer environment database provisioned for People Depot developers to work against.
### Action Items
- [ ] Provision PostgresQL RDS
- [ ] Give dev credentials to People Depot
### Resources/Instructions
https://repost.aws/knowledge-center/ecs-fargate-task-database-connection
| reli | provision rds database for people depot dev environment overview as a people depot developer i d like an developer environment database provisioned for people depot developers to work against action items provision postgresql rds give dev credentials to people depot resources instructions | 1 |
1,109 | 13,146,873,446 | IssuesEvent | 2020-08-08 12:33:03 | osbuild/osbuild-composer | https://api.github.com/repos/osbuild/osbuild-composer | opened | ci: exceeding openstack quota | ci-reliability | Now that our image tests spawn openstack instances we risk exceeding our quota, leading to test failure.
A simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need. | True | ci: exceeding openstack quota - Now that our image tests spawn openstack instances we risk exceeding our quota, leading to test failure.
A simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need. | reli | ci exceeding openstack quota now that our image tests spawn openstack instances we risk exceeding our quota leading to test failure a simple solution to this would be to limit the number of parallel tests we run to account for how many openstack instances each test may need | 1 |
110,010 | 16,963,463,390 | IssuesEvent | 2021-06-29 08:03:47 | elastic/kibana | https://api.github.com/repos/elastic/kibana | reopened | [Security Solution] Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West | QA:Validated Team: SecuritySolution Team:Threat Hunting bug v7.13.0 | **Describe the feature**
Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West
**Build Details:**
```
Version: 7.12.1 BC3
Commit: d03a18355d794e508183a34734141626ce74b2c4
Build: 39452
Artifact Page: https://staging.elastic.co/7.12.1-7947c5bf/summary-7.12.1.html
```
**Preconditions**
1. Elastic 7.12.1 environment should be deployed
2. Alerts/Events should be generated
**Steps to Reproduce**
1. Navigate to the Hosts Tab.
2. Go to the External alerts and open the analyzer for any alert/event
3. Observe that proper UI shows with nodes and information on the UI
4. Click on the North, South, East or West arrows until the node crosses the extremes of the UI allowed for the UI
5. Observe that the UI starts showing the message 'No Process Events Found'.
**Test data**
N/A
**Impacted Test case(s)**
N/A
**Actual Result**
Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West
**Expected Result**
Analyzer should show the nodes and data even if the user navigates to extreme North, South, East or West
**What's Working**
N/A
**What's Not Working**
N/A
**Screenshot**
- Normal analyzer

- When the user navigates to extreme North, South, East or West
 | True | [Security Solution] Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West - **Describe the feature**
Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West
**Build Details:**
```
Version: 7.12.1 BC3
Commit: d03a18355d794e508183a34734141626ce74b2c4
Build: 39452
Artifact Page: https://staging.elastic.co/7.12.1-7947c5bf/summary-7.12.1.html
```
**Preconditions**
1. Elastic 7.12.1 environment should be deployed
2. Alerts/Events should be generated
**Steps to Reproduce**
1. Navigate to the Hosts Tab.
2. Go to the External alerts and open the analyzer for any alert/event
3. Observe that proper UI shows with nodes and information on the UI
4. Click on the North, South, East or West arrows until the node crosses the extremes of the UI allowed for the UI
5. Observe that the UI starts showing the message 'No Process Events Found'.
**Test data**
N/A
**Impacted Test case(s)**
N/A
**Actual Result**
Analyzer UI shows the message 'No Process Events Found' if the user navigates to extreme North, South, East or West
**Expected Result**
Analyzer should show the nodes and data even if the user navigates to extreme North, South, East or West
**What's Working**
N/A
**What's Not Working**
N/A
**Screenshot**
- Normal analyzer

- When the user navigates to extreme North, South, East or West
 | non_reli | analyzer ui shows the message no process events found if the user navigates to extreme north south east or west describe the feature analyzer ui shows the message no process events found if the user navigates to extreme north south east or west build details version commit build artifact page preconditions elastic environment should be deployed alerts events should be generated steps to reproduce navigate to the hosts tab go to the external alerts and open the analyzer for any alert event observe that proper ui shows with nodes and information on the ui click on the north south east or west arrows until the node crosses the extremes of the ui allowed for the ui observe that the ui starts showing the message no process events found test data n a impacted test case s n a actual result analyzer ui shows the message no process events found if the user navigates to extreme north south east or west expected result analyzer should show the nodes and data even if the user navigates to extreme north south east or west what s working n a what s not working n a screenshot normal analyzer when the user navigates to extreme north south east or west | 0 |
1,559 | 17,064,819,271 | IssuesEvent | 2021-07-07 05:32:50 | MicrosoftDocs/live-share | https://api.github.com/repos/MicrosoftDocs/live-share | closed | Infinite loading and disconnecting issues | needs more info product-feedback reliability | Hello,
Me and my team have been using this product for more than half a year now, but more than ever, it seems to be glitching everywhere and generally causing a lot of problems.
Currently, we experience the following issues: changes don't save often times, the localhost is randomly failing, plus infinite or generally extremely slow loading speeds when simply changing a file or searching (file searching is a complete mess, it merely worked before).
Not sure what we can do about this, since it's been increasingly frustrating the past couple of weeks, and it's not related to our server or our connection, which is through Wireless, same as before. Any help or advice would be welcome. Thanks! | True | Infinite loading and disconnecting issues - Hello,
Me and my team have been using this product for more than half a year now, but more than ever, it seems to be glitching everywhere and generally causing a lot of problems.
Currently, we experience the following issues: changes don't save often times, the localhost is randomly failing, plus infinite or generally extremely slow loading speeds when simply changing a file or searching (file searching is a complete mess, it merely worked before).
Not sure what we can do about this, since it's been increasingly frustrating the past couple of weeks, and it's not related to our server or our connection, which is through Wireless, same as before. Any help or advice would be welcome. Thanks! | reli | infinite loading and disconnecting issues hello me and my team have been using this product for more than half a year now but more than ever it seems to be glitching everywhere and generally causing a lot of problems currently we experience the following issues changes don t save often times the localhost is randomly failing plus infinite or generally extremely slow loading speeds when simply changing a file or searching file searching is a complete mess it merely worked before not sure what we can do about this since it s been increasingly frustrating the past couple of weeks and it s not related to our server or our connection which is through wireless same as before any help or advice would be welcome thanks | 1 |
752 | 10,348,015,677 | IssuesEvent | 2019-09-04 18:46:02 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Cancelling Http requests can result in "The handle is invalid" or other WinHTTP state errors | area-System.Net.Http bug os-windows tenet-reliability | Http requests can be cancelled in various ways using a cancellation token or even a timeout. Sometimes the CI tests timeout contacting the Azure test server due to network congestion, so this results in the requests getting cancelled as well by the HttpClient.Timeout property.
When Http requests are cancelled in WinHttpHandler, the WinHTTP handle is closed. This cancels any outstanding async operations in WinHTTP. However, the current CancellationTokenRegistration code is closing the handle without first seeing if there is an active operation in flight.
``` c#
private async void StartRequest(object obj)
{
// ...
using (state.CancellationToken.Register(s => ((WinHttpRequestState)s).RequestHandle.Dispose(), state))
{
do
{
// ...
} while (state.RetryRequest);
}
```
Since this cancellation is happening on a different thread from the main StartRequest() logic, it results in in-deterministic behavior.
PR #8693 (Fix cancellation of WinHttpHandler response stream reads) has a similar CancellationToken.Register logic but it correctly checks for an active operation. This same logic needs to apply to the StartRequest() operation.
This is the root cause of the following issues:
#8928
#8496
#8426
#8061
#8000
## History of failures
```
System.Net.Http.HttpRequestException : An error occurred while sending the request.
---- System.Net.Http.WinHttpException : The handle is invalid
```
Day | Build | OS | Test
-- | -- | -- | --
5/9 | 20170509.01 | Win7 | Timeout_SetTo30AndGetResponseFromLoopbackQuickly_Success
5/10 | 20170510.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
5/11 | 20170511.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
5/19 | 20170519.04 | Win7 | PostAsync_CallMethod_StreamContent
5/26 | 20170526.01 | Win7 | PostLongerContentLengths_UsesChunkedSemantics
5/30 | 20170530.06 | Win7 | SetDelegate_ConnectionSucceeds
6/4 | 20170604.02 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds
6/7 | 20170607.01 | Win7 | PostSyncBlockingContentUsingChunkedEncoding_Success
6/13 | 20170613.02 | Win7 | SetDelegate_ConnectionSucceeds
6/14 | 20170614.01 | Win7 | PostAsync_CallMethod_StreamContent
6/15 | 20170615.03 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds
6/15 | 20170615.03 | Win8.1 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized
6/16 | 20170616.01 | Win7 | PostAsync_CallMethod_EmptyContent
9/26 | 20170926.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
10/10 | 20171010.03 | Win7 | GetStreamAsync_ReadToEnd_Success
10/11 | 20171011.02 | Win7 | PostAsync_CallMethod_StreamContent
10/20 | 20171020.01 | Win7 | SetDelegate_ConnectionSucceeds
10/22 | 20171022.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent
10/22 | 20171022.02 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent
10/23 | 20171023.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
10/25 | 20171025.04 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
10/25 | 20171025.08 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
10/25 | 20171025.08 | Win7 | PostAsync_CallMethod_StreamContent
10/27 | 20171027.03 | Win7 | PostUsingContentLengthSemantics_Success
10/30 | 20171030.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
10/30 | 20171030.02 | Win7 | GetAsync_SupportedSSLVersion_Succeeds
10/31 | 20171030.05 | Win7 | SetDelegate_ConnectionSucceeds
11/3 | 20171103.07 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/6 | 20171106.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/6 | 20171106.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
11/7 | 20171107.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/9 | 20171109.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
11/15 | 20171115.02 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent
11/15 | 20171115.03 | Win7 | GetAsync_TrailingHeaders_Ignored
11/18 | 20171118.04 | Win7 | GetStreamAsync_ReadZeroBytes_Success
11/20 | 20171120.01 | Win7 | GetAsync_TrailingHeaders_Ignored
11/20 | 20171120.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
11/22 | 20171122.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/23 | 20171123.02 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized
11/24 | 20171124.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/27 | 20171127.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/28 | 20171128.01 | Win7 | SetDelegate_ConnectionSucceeds
11/28 | 20171128.07 | Win10 | UseCallback_NotSecureConnection_CallbackNotCalled
12/5 | 20171205.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
12/5 | 20171205.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
12/8 | 20171208.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent
12/12 | 20171212.02 | Win7 | GetStreamAsync_ReadToEnd_Success
12/23 | 20171223.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171223.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
12/24 | 20171224.02 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171224.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
12/31 | 20171231.01 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20171231.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized(authHeaders:%20%5C%22WWW-Authenticate:%20CustomAuth%5C%5Cr%5C%5Cn%5C%22))
1/5 | 20180105.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180105.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True))
1/10 | 20180110.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180110.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20201%20Created%5C%22,%20expectedStatusCode:%20201,%20expectedReason:%20%5C%22Created%5C%22))
1/11 | 20180111.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180111.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20500%20Internal%20Server%20Error%5C%22,%20expectedStatusCode:%20500,%20expectedReason:%20%5C%22Inter))
1/14 | 20180114.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180114.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostAsync_EmptyContent_ContentTypeHeaderNotSent(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/16 | 20180116.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180116.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20True))
1/18 | 20180118.04 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180118.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False))
1/19 | 20180119.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20180119.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test%7E2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20False))
1/19 | 20180119.04 | Win7 | PostUsingContentLengthSemantics_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180119.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingContentLengthSemantics_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/21 | 20180121.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180121.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False))
1/22 | 20180122.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180122.02/workItem/System.Net.Http.Functional.Tests/wilogs)
1/23 | 20180123.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRewindableContentUsingAuth_NoPreAuthenticate_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx%3Fauth=basic&user=user1&password=password1))
1/23 | 20180123.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22))
1/23 | 20180123.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
1/25 | 20180125.01 | Win7 | PostAsync_ReuseRequestContent_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180125.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FPostAsync_ReuseRequestContent_Success(remoteServer:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/27 | 20180127.01 | Win7 | GetAsync_TrailingHeaders_Ignored - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_TrailingHeaders_Ignored(includeTrailerHeader:%20True))
1/27 | 20180127.04 | Win10 | PostRepeatedFlushContentUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRepeatedFlushContentUsingChunkedEncoding_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/27 | 20180127.04 | Win7 | GetAsync_RedirectResponseHasCookie_CookieSentToFinalUri - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/wilogs)
1/29 | 20180129.05 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180129.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
1/30 | 20180130.05 | Win7 | PostEmptyContentUsingConflictingSemantics_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180130.05/workItem/System.Net.Http.Functional.Tests/wilogs)
1/31 | 20180131.05 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180131.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True))
2/1 | 20180201.01 | Win7 | PostUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingChunkedEncoding_Success(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
2/1 | 20180201.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
2/1 | 20180201.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22))
2/5 | 20180205.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180205.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False))
2/6 | 20180206.04 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False))
2/6 | 20180206.05 | Win7 | UseCallback_NotSecureConnection_CallbackNotCalled - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_ServerCertificates_Test~2FUseCallback_NotSecureConnection_CallbackNotCalled)
2/8 | 20180208.01 | Win7 | 2x GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20O%20K%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22O%20K%5C%22)) and [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20600%20still%20valid%5C%22,%20expectedStatusCode:%20600,%20expectedReason:%20%5C%22still%20valid%5C%22))
2/9 | 20180209.05 | Win7 | ThresholdExceeded_ThrowsException - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180209.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxResponseHeadersLength_Test~2FThresholdExceeded_ThrowsException(responseHeaders:%20%5C%22HTTP~2F1.1%20200%20OK%5C%5Cr%5C%5CnContent-Length:%200%5C%5Cr%5C%5CnCustom-000%5C%22...,%20maxResponseHeadersLe))
2/10 | 20180210.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
2/10 | 20180210.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/wilogs)
2/13 | 20180213.01 | Win7 | ThresholdExceeded_ThrowsException
2/20 | 20180220.08 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success
2/21 | 20180221.01 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success
2/25 | 20180225.01 | Win7 | HttpClientHandler_IncorrectCredentials_Fails - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_IncorrectCredentials_Fails(authenticateHeader:%20%5C%22WWW-Authenticate:%20Digest%20realm=%5C%5C%5C%22hello%5C%5C%5C%22,%20nonce=%5C%5C%5C%22%5C%22...))
2/25 | 20180225.02 | Win10 | HttpClientHandler_Authentication_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_Authentication_Succeeds(authenticateHeader:%20%5C%22Basic%20something,%20Digest%20something%5C%22,%20result:%20False))
2/26 | 20180226.04 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180226.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test~2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20True))
2/28 | 20180228.07 | Win7 | ThresholdExceeded_ThrowsException
3/1 | 20180301.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/2 | 20180302.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/4 | 20180304.01 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/7 | 20180307.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed
3/7 | 20180307.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/7 | 20180307.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/8 | 20180308.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/8 | 20180308.09 | Win7 | GetAsync_SetCookieContainer_CookieSent
3/10 | 20180310.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/14 | 20180314.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/17 | 20180317.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed
3/22 | 20180322.06 | Win7 | ThresholdExceeded_ThrowsException
#### Similar failure with IOException wrapped in the middle
```
System.Net.Http.HttpRequestException : Error while copying content to a stream.
---- System.IO.IOException : The write operation failed, see inner exception.
-------- System.Net.Http.WinHttpException : The handle is invalid
```
Day | Build | OS | Test
-- | -- | -- | --
11/17 | 20171117.04 | Win7 | PostUsingContentLengthSemantics_Success
12/2 | 20171202.01 | Win7 | PostAsync_ReuseRequestContent_Success | True | Cancelling Http requests can result in "The handle is invalid" or other WinHTTP state errors - Http requests can be cancelled in various ways using a cancellation token or even a timeout. Sometimes the CI tests timeout contacting the Azure test server due to network congestion, so this results in the requests getting cancelled as well by the HttpClient.Timeout property.
When Http requests are cancelled in WinHttpHandler, the WinHTTP handle is closed. This cancels any outstanding async operations in WinHTTP. However, the current CancellationTokenRegistration code is closing the handle without first seeing if there is an active operation in flight.
``` c#
private async void StartRequest(object obj)
{
// ...
using (state.CancellationToken.Register(s => ((WinHttpRequestState)s).RequestHandle.Dispose(), state))
{
do
{
// ...
} while (state.RetryRequest);
}
```
Since this cancellation is happening on a different thread from the main StartRequest() logic, it results in in-deterministic behavior.
PR #8693 (Fix cancellation of WinHttpHandler response stream reads) has a similar CancellationToken.Register logic but it correctly checks for an active operation. This same logic needs to apply to the StartRequest() operation.
This is the root cause of the following issues:
#8928
#8496
#8426
#8061
#8000
## History of failures
```
System.Net.Http.HttpRequestException : An error occurred while sending the request.
---- System.Net.Http.WinHttpException : The handle is invalid
```
Day | Build | OS | Test
-- | -- | -- | --
5/9 | 20170509.01 | Win7 | Timeout_SetTo30AndGetResponseFromLoopbackQuickly_Success
5/10 | 20170510.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
5/11 | 20170511.01 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
5/19 | 20170519.04 | Win7 | PostAsync_CallMethod_StreamContent
5/26 | 20170526.01 | Win7 | PostLongerContentLengths_UsesChunkedSemantics
5/30 | 20170530.06 | Win7 | SetDelegate_ConnectionSucceeds
6/4 | 20170604.02 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds
6/7 | 20170607.01 | Win7 | PostSyncBlockingContentUsingChunkedEncoding_Success
6/13 | 20170613.02 | Win7 | SetDelegate_ConnectionSucceeds
6/14 | 20170614.01 | Win7 | PostAsync_CallMethod_StreamContent
6/15 | 20170615.03 | Win7 | NoCallback_RevokedCertificate_NoRevocationChecking_Succeeds
6/15 | 20170615.03 | Win8.1 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized
6/16 | 20170616.01 | Win7 | PostAsync_CallMethod_EmptyContent
9/26 | 20170926.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
10/10 | 20171010.03 | Win7 | GetStreamAsync_ReadToEnd_Success
10/11 | 20171011.02 | Win7 | PostAsync_CallMethod_StreamContent
10/20 | 20171020.01 | Win7 | SetDelegate_ConnectionSucceeds
10/22 | 20171022.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent
10/22 | 20171022.02 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent
10/23 | 20171023.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
10/25 | 20171025.04 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
10/25 | 20171025.08 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
10/25 | 20171025.08 | Win7 | PostAsync_CallMethod_StreamContent
10/27 | 20171027.03 | Win7 | PostUsingContentLengthSemantics_Success
10/30 | 20171030.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
10/30 | 20171030.02 | Win7 | GetAsync_SupportedSSLVersion_Succeeds
10/31 | 20171030.05 | Win7 | SetDelegate_ConnectionSucceeds
11/3 | 20171103.07 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/6 | 20171106.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/6 | 20171106.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success
11/7 | 20171107.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/9 | 20171109.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
11/15 | 20171115.02 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent
11/15 | 20171115.03 | Win7 | GetAsync_TrailingHeaders_Ignored
11/18 | 20171118.04 | Win7 | GetStreamAsync_ReadZeroBytes_Success
11/20 | 20171120.01 | Win7 | GetAsync_TrailingHeaders_Ignored
11/20 | 20171120.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds
11/22 | 20171122.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/23 | 20171123.02 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized
11/24 | 20171124.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/27 | 20171127.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
11/28 | 20171128.01 | Win7 | SetDelegate_ConnectionSucceeds
11/28 | 20171128.07 | Win10 | UseCallback_NotSecureConnection_CallbackNotCalled
12/5 | 20171205.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
12/5 | 20171205.06 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
12/8 | 20171208.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent
12/12 | 20171212.02 | Win7 | GetStreamAsync_ReadToEnd_Success
12/23 | 20171223.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171223.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
12/24 | 20171224.02 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20171224.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test%7E2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
12/31 | 20171231.01 | Win7 | GetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20171231.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_ServerNeedsNonStandardAuthAndSetCredential_StatusCodeUnauthorized(authHeaders:%20%5C%22WWW-Authenticate:%20CustomAuth%5C%5Cr%5C%5Cn%5C%22))
1/5 | 20180105.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180105.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True))
1/10 | 20180110.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180110.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20201%20Created%5C%22,%20expectedStatusCode:%20201,%20expectedReason:%20%5C%22Created%5C%22))
1/11 | 20180111.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180111.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests_Dribble~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20500%20Internal%20Server%20Error%5C%22,%20expectedStatusCode:%20500,%20expectedReason:%20%5C%22Inter))
1/14 | 20180114.01 | Win7 | PostAsync_EmptyContent_ContentTypeHeaderNotSent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180114.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostAsync_EmptyContent_ContentTypeHeaderNotSent(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/16 | 20180116.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180116.03/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20True))
1/18 | 20180118.04 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180118.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False))
1/19 | 20180119.02 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official%7E2Fcorefx%7E2Fmaster%7E2F/type/test%7E2Ffunctional%7E2Fcli%7E2F/build/20180119.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test%7E2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20False))
1/19 | 20180119.04 | Win7 | PostUsingContentLengthSemantics_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180119.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingContentLengthSemantics_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/21 | 20180121.01 | Win7 | MaxResponseContentBufferSize_ThrowsIfTooSmallForContent - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180121.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FMaxResponseContentBufferSize_ThrowsIfTooSmallForContent(maxSize:%202,%20contentLength:%201,%20exceptionExpected:%20False))
1/22 | 20180122.02 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180122.02/workItem/System.Net.Http.Functional.Tests/wilogs)
1/23 | 20180123.01 | Win7 | PostRewindableContentUsingAuth_NoPreAuthenticate_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRewindableContentUsingAuth_NoPreAuthenticate_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx%3Fauth=basic&user=user1&password=password1))
1/23 | 20180123.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22))
1/23 | 20180123.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180123.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
1/25 | 20180125.01 | Win7 | PostAsync_ReuseRequestContent_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180125.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FPostAsync_ReuseRequestContent_Success(remoteServer:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/27 | 20180127.01 | Win7 | GetAsync_TrailingHeaders_Ignored - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FGetAsync_TrailingHeaders_Ignored(includeTrailerHeader:%20True))
1/27 | 20180127.04 | Win10 | PostRepeatedFlushContentUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostRepeatedFlushContentUsingChunkedEncoding_Success(serverUri:%20http:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
1/27 | 20180127.04 | Win7 | GetAsync_RedirectResponseHasCookie_CookieSentToFinalUri - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180127.04/workItem/System.Net.Http.Functional.Tests/wilogs)
1/29 | 20180129.05 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180129.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%202,%20secure:%20True))
1/30 | 20180130.05 | Win7 | PostEmptyContentUsingConflictingSemantics_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180130.05/workItem/System.Net.Http.Functional.Tests/wilogs)
1/31 | 20180131.05 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180131.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20Tls,%20Tls11,%20Tls12,%20requestOnlyThisProtocol:%20True))
2/1 | 20180201.01 | Win7 | PostUsingChunkedEncoding_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.PostScenarioTest~2FPostUsingChunkedEncoding_Success(serverUri:%20https:~2F~2Fcorefx-net.cloudapp.net~2FEcho.ashx))
2/1 | 20180201.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
2/1 | 20180201.05 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180201.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20OK%C2%80%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22OK%3F%5C%22))
2/5 | 20180205.01 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180205.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False))
2/6 | 20180206.04 | Win7 | SetDelegate_ConnectionSucceeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_DangerousAcceptAllCertificatesValidator_Test~2FSetDelegate_ConnectionSucceeds(acceptedProtocol:%20None,%20requestOnlyThisProtocol:%20False))
2/6 | 20180206.05 | Win7 | UseCallback_NotSecureConnection_CallbackNotCalled - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180206.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_ServerCertificates_Test~2FUseCallback_NotSecureConnection_CallbackNotCalled)
2/8 | 20180208.01 | Win7 | 2x GetAsync_ExpectedStatusCodeAndReason_Success - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20200%20O%20K%5C%22,%20expectedStatusCode:%20200,%20expectedReason:%20%5C%22O%20K%5C%22)) and [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180208.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpProtocolTests~2FGetAsync_ExpectedStatusCodeAndReason_Success(statusLine:%20%5C%22HTTP~2F1.1%20600%20still%20valid%5C%22,%20expectedStatusCode:%20600,%20expectedReason:%20%5C%22still%20valid%5C%22))
2/9 | 20180209.05 | Win7 | ThresholdExceeded_ThrowsException - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180209.05/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxResponseHeadersLength_Test~2FThresholdExceeded_ThrowsException(responseHeaders:%20%5C%22HTTP~2F1.1%20200%20OK%5C%5Cr%5C%5CnContent-Length:%200%5C%5Cr%5C%5CnCustom-000%5C%22...,%20maxResponseHeadersLe))
2/10 | 20180210.01 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_MaxConnectionsPerServer_Test~2FGetAsync_MaxLimited_ConcurrentCallsStillSucceed(maxConnections:%203,%20numRequests:%205,%20secure:%20False))
2/10 | 20180210.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success - only in [catastrophic failure logs](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180210.01/workItem/System.Net.Http.Functional.Tests/wilogs)
2/13 | 20180213.01 | Win7 | ThresholdExceeded_ThrowsException
2/20 | 20180220.08 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success
2/21 | 20180221.01 | Win7 | The handle is invalid | GetAsync_ExpectedStatusCodeAndReason_Success
2/25 | 20180225.01 | Win7 | HttpClientHandler_IncorrectCredentials_Fails - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_IncorrectCredentials_Fails(authenticateHeader:%20%5C%22WWW-Authenticate:%20Digest%20realm=%5C%5C%5C%22hello%5C%5C%5C%22,%20nonce=%5C%5C%5C%22%5C%22...))
2/25 | 20180225.02 | Win10 | HttpClientHandler_Authentication_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180225.02/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_Authentication_Test~2FHttpClientHandler_Authentication_Succeeds(authenticateHeader:%20%5C%22Basic%20something,%20Digest%20something%5C%22,%20result:%20False))
2/26 | 20180226.04 | Win7 | GetAsync_AllowedSSLVersion_Succeeds - [link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180226.04/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandler_SslProtocols_Test~2FGetAsync_AllowedSSLVersion_Succeeds(acceptedProtocol:%20Tls11,%20requestOnlyThisProtocol:%20True))
2/28 | 20180228.07 | Win7 | ThresholdExceeded_ThrowsException
3/1 | 20180301.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/2 | 20180302.06 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/4 | 20180304.01 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/7 | 20180307.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed
3/7 | 20180307.03 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/7 | 20180307.04 | Win7 | ReadAsStreamAsync_ValidServerResponse_Success
3/8 | 20180308.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/8 | 20180308.09 | Win7 | GetAsync_SetCookieContainer_CookieSent
3/10 | 20180310.02 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/14 | 20180314.01 | Win7 | GetAsync_ExpectedStatusCodeAndReason_Success
3/17 | 20180317.03 | Win7 | GetAsync_MaxLimited_ConcurrentCallsStillSucceed
3/22 | 20180322.06 | Win7 | ThresholdExceeded_ThrowsException
#### Similar failure with IOException wrapped in the middle
```
System.Net.Http.HttpRequestException : Error while copying content to a stream.
---- System.IO.IOException : The write operation failed, see inner exception.
-------- System.Net.Http.WinHttpException : The handle is invalid
```
Day | Build | OS | Test
-- | -- | -- | --
11/17 | 20171117.04 | Win7 | PostUsingContentLengthSemantics_Success
12/2 | 20171202.01 | Win7 | PostAsync_ReuseRequestContent_Success | reli | cancelling http requests can result in the handle is invalid or other winhttp state errors http requests can be cancelled in various ways using a cancellation token or even a timeout sometimes the ci tests timeout contacting the azure test server due to network congestion so this results in the requests getting cancelled as well by the httpclient timeout property when http requests are cancelled in winhttphandler the winhttp handle is closed this cancels any outstanding async operations in winhttp however the current cancellationtokenregistration code is closing the handle without first seeing if there is an active operation in flight c private async void startrequest object obj using state cancellationtoken register s winhttprequeststate s requesthandle dispose state do while state retryrequest since this cancellation is happening on a different thread from the main startrequest logic it results in in deterministic behavior pr fix cancellation of winhttphandler response stream reads has a similar cancellationtoken register logic but it correctly checks for an active operation this same logic needs to apply to the startrequest operation this is the root cause of the following issues history of failures system net http httprequestexception an error occurred while sending the request system net http winhttpexception the handle is invalid day build os test timeout success getasync allowedsslversion succeeds getasync allowedsslversion succeeds postasync callmethod streamcontent postlongercontentlengths useschunkedsemantics setdelegate connectionsucceeds nocallback revokedcertificate norevocationchecking succeeds postsyncblockingcontentusingchunkedencoding success setdelegate connectionsucceeds postasync callmethod streamcontent nocallback revokedcertificate norevocationchecking succeeds getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized postasync callmethod emptycontent readasstreamasync validserverresponse success getstreamasync readtoend success postasync callmethod streamcontent setdelegate connectionsucceeds maxresponsecontentbuffersize throwsiftoosmallforcontent maxresponsecontentbuffersize throwsiftoosmallforcontent readasstreamasync validserverresponse success getasync expectedstatuscodeandreason success postrewindablecontentusingauth nopreauthenticate success postasync callmethod streamcontent postusingcontentlengthsemantics success postrewindablecontentusingauth nopreauthenticate success getasync supportedsslversion succeeds setdelegate connectionsucceeds getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postrewindablecontentusingauth nopreauthenticate success getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success postasync emptycontent contenttypeheadernotsent getasync trailingheaders ignored getstreamasync readzerobytes success getasync trailingheaders ignored getasync allowedsslversion succeeds getasync expectedstatuscodeandreason success getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success setdelegate connectionsucceeds usecallback notsecureconnection callbacknotcalled getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postasync emptycontent contenttypeheadernotsent getstreamasync readtoend success getasync maxlimited concurrentcallsstillsucceed getasync maxlimited concurrentcallsstillsucceed getasync serverneedsnonstandardauthandsetcredential statuscodeunauthorized setdelegate connectionsucceeds getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success postasync emptycontent contenttypeheadernotsent setdelegate connectionsucceeds maxresponsecontentbuffersize throwsiftoosmallforcontent getasync allowedsslversion succeeds postusingcontentlengthsemantics success maxresponsecontentbuffersize throwsiftoosmallforcontent readasstreamasync validserverresponse success only in postrewindablecontentusingauth nopreauthenticate success getasync expectedstatuscodeandreason success getasync maxlimited concurrentcallsstillsucceed postasync reuserequestcontent success getasync trailingheaders ignored postrepeatedflushcontentusingchunkedencoding success getasync redirectresponsehascookie cookiesenttofinaluri only in getasync maxlimited concurrentcallsstillsucceed postemptycontentusingconflictingsemantics success only in setdelegate connectionsucceeds postusingchunkedencoding success getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success setdelegate connectionsucceeds setdelegate connectionsucceeds usecallback notsecureconnection callbacknotcalled getasync expectedstatuscodeandreason success and thresholdexceeded throwsexception getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success only in thresholdexceeded throwsexception the handle is invalid getasync expectedstatuscodeandreason success the handle is invalid getasync expectedstatuscodeandreason success httpclienthandler incorrectcredentials fails httpclienthandler authentication succeeds getasync allowedsslversion succeeds thresholdexceeded throwsexception getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success readasstreamasync validserverresponse success getasync maxlimited concurrentcallsstillsucceed getasync expectedstatuscodeandreason success readasstreamasync validserverresponse success getasync expectedstatuscodeandreason success getasync setcookiecontainer cookiesent getasync expectedstatuscodeandreason success getasync expectedstatuscodeandreason success getasync maxlimited concurrentcallsstillsucceed thresholdexceeded throwsexception similar failure with ioexception wrapped in the middle system net http httprequestexception error while copying content to a stream system io ioexception the write operation failed see inner exception system net http winhttpexception the handle is invalid day build os test postusingcontentlengthsemantics success postasync reuserequestcontent success | 1 |
5,657 | 3,972,533,487 | IssuesEvent | 2016-05-04 15:36:21 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Mounting an AWS volume returns an error that looks incorrect: Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume | area/usability component/storage | This looks like an incorrect setup. Is this the dynamic provisioner? The error message is not useful to admins or users.
```
- apiVersion: v1
count: 1
firstTimestamp: 2016-05-04T14:33:04Z
involvedObject:
apiVersion: v1
kind: Pod
name: jenkins-1-c1629
namespace: jwforres
resourceVersion: "3983188"
uid: fcddad4f-1204-11e6-98bd-0a1d348c34bb
kind: Event
lastTimestamp: 2016-05-04T14:33:04Z
message: 'Unable to mount volumes for pod "jenkins-1-c1629_jwforres(fcddad4f-1204-11e6-98bd-0a1d348c34bb)":
Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS
volume (aws:///placeholder-for-provisioning)'
metadata:
creationTimestamp: 2016-05-04T14:33:04Z
deletionTimestamp: 2016-05-04T16:33:04Z
name: jenkins-1-c1629.144b63ef823f1dc1
namespace: jwforres
resourceVersion: "3983383"
selfLink: /api/v1/namespaces/jwforres/events/jenkins-1-c1629.144b63ef823f1dc1
uid: 1c7cf10f-1205-11e6-abb2-0ada84b8265d
reason: FailedMount
source:
component: kubelet
host: ip-172-31-14-20.ec2.internal
type: Warning
``` | True | Mounting an AWS volume returns an error that looks incorrect: Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS volume - This looks like an incorrect setup. Is this the dynamic provisioner? The error message is not useful to admins or users.
```
- apiVersion: v1
count: 1
firstTimestamp: 2016-05-04T14:33:04Z
involvedObject:
apiVersion: v1
kind: Pod
name: jenkins-1-c1629
namespace: jwforres
resourceVersion: "3983188"
uid: fcddad4f-1204-11e6-98bd-0a1d348c34bb
kind: Event
lastTimestamp: 2016-05-04T14:33:04Z
message: 'Unable to mount volumes for pod "jenkins-1-c1629_jwforres(fcddad4f-1204-11e6-98bd-0a1d348c34bb)":
Could not attach EBS Disk "placeholder-for-provisioning": Invalid format for AWS
volume (aws:///placeholder-for-provisioning)'
metadata:
creationTimestamp: 2016-05-04T14:33:04Z
deletionTimestamp: 2016-05-04T16:33:04Z
name: jenkins-1-c1629.144b63ef823f1dc1
namespace: jwforres
resourceVersion: "3983383"
selfLink: /api/v1/namespaces/jwforres/events/jenkins-1-c1629.144b63ef823f1dc1
uid: 1c7cf10f-1205-11e6-abb2-0ada84b8265d
reason: FailedMount
source:
component: kubelet
host: ip-172-31-14-20.ec2.internal
type: Warning
``` | non_reli | mounting an aws volume returns an error that looks incorrect could not attach ebs disk placeholder for provisioning invalid format for aws volume this looks like an incorrect setup is this the dynamic provisioner the error message is not useful to admins or users apiversion count firsttimestamp involvedobject apiversion kind pod name jenkins namespace jwforres resourceversion uid kind event lasttimestamp message unable to mount volumes for pod jenkins jwforres could not attach ebs disk placeholder for provisioning invalid format for aws volume aws placeholder for provisioning metadata creationtimestamp deletiontimestamp name jenkins namespace jwforres resourceversion selflink api namespaces jwforres events jenkins uid reason failedmount source component kubelet host ip internal type warning | 0 |
2,932 | 30,316,765,236 | IssuesEvent | 2023-07-10 16:05:20 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | closed | Cancel on-going remote stream registration on stream removal | kind/bug area/performance area/reliability component/transport | **Describe the bug**
There is currently a potential race condition which would result in a remote stream existing server side, even though the client stream has gone away.
Since we register remote streams asynchronously, a remove request may be submitted client side, which will immediately remove it there. Then asynchronous removal requests are sent to the server. However, this can be interleaved with the asynchronous registration, resulting in a stream existing server side.
The impact is additional latency during a push, or possible unnecessary job activation if it was the last stream for this type. However, the stream will eventually get removed appropriately.
**Expected behavior**
Registration/removal of remote streams is sequenced, such that a removal request would cancel registration attempts, and queue the removal after whatever in-flight requests were sent are finished.
There is still a slight edge case around time outs, of course, but I think this is acceptable for now. The other option would be introducing even more coordination in the protocol, and I'd rather avoid this. | True | Cancel on-going remote stream registration on stream removal - **Describe the bug**
There is currently a potential race condition which would result in a remote stream existing server side, even though the client stream has gone away.
Since we register remote streams asynchronously, a remove request may be submitted client side, which will immediately remove it there. Then asynchronous removal requests are sent to the server. However, this can be interleaved with the asynchronous registration, resulting in a stream existing server side.
The impact is additional latency during a push, or possible unnecessary job activation if it was the last stream for this type. However, the stream will eventually get removed appropriately.
**Expected behavior**
Registration/removal of remote streams is sequenced, such that a removal request would cancel registration attempts, and queue the removal after whatever in-flight requests were sent are finished.
There is still a slight edge case around time outs, of course, but I think this is acceptable for now. The other option would be introducing even more coordination in the protocol, and I'd rather avoid this. | reli | cancel on going remote stream registration on stream removal describe the bug there is currently a potential race condition which would result in a remote stream existing server side even though the client stream has gone away since we register remote streams asynchronously a remove request may be submitted client side which will immediately remove it there then asynchronous removal requests are sent to the server however this can be interleaved with the asynchronous registration resulting in a stream existing server side the impact is additional latency during a push or possible unnecessary job activation if it was the last stream for this type however the stream will eventually get removed appropriately expected behavior registration removal of remote streams is sequenced such that a removal request would cancel registration attempts and queue the removal after whatever in flight requests were sent are finished there is still a slight edge case around time outs of course but i think this is acceptable for now the other option would be introducing even more coordination in the protocol and i d rather avoid this | 1 |
468,217 | 13,463,030,587 | IssuesEvent | 2020-09-09 16:56:40 | googleapis/releasetool | https://api.github.com/repos/googleapis/releasetool | closed | Magic proxy is currently failing, resulting in labels not being removed post publication | priority: p2 type: bug | #### Steps to reproduce
1. Merge a release PR, resulting in autorelease tagging a release.
2. It will enqueue a job which kicks off publication on kokoro.
3. When this job finishes, publication will succeed, but labels will not be released; the following error will be in logs:
```bash
requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: https://magic-github-proxy.endpoints.devrel-prod.cloud.goog/repos/googleapis/java-document-ai/issues/133/comments?
```
| 1.0 | Magic proxy is currently failing, resulting in labels not being removed post publication - #### Steps to reproduce
1. Merge a release PR, resulting in autorelease tagging a release.
2. It will enqueue a job which kicks off publication on kokoro.
3. When this job finishes, publication will succeed, but labels will not be released; the following error will be in logs:
```bash
requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: https://magic-github-proxy.endpoints.devrel-prod.cloud.goog/repos/googleapis/java-document-ai/issues/133/comments?
```
| non_reli | magic proxy is currently failing resulting in labels not being removed post publication steps to reproduce merge a release pr resulting in autorelease tagging a release it will enqueue a job which kicks off publication on kokoro when this job finishes publication will succeed but labels will not be released the following error will be in logs bash requests exceptions httperror client error bad request for url | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.