Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
9,069
| 12,140,104,440
|
IssuesEvent
|
2020-04-23 19:58:13
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
remove gcp-devrel-py-tools from monitoring/api/v3/cloud-client/requirements-test.txt
|
priority: p2 remove-gcp-devrel-py-tools type: process
|
remove gcp-devrel-py-tools from monitoring/api/v3/cloud-client/requirements-test.txt
|
1.0
|
remove gcp-devrel-py-tools from monitoring/api/v3/cloud-client/requirements-test.txt - remove gcp-devrel-py-tools from monitoring/api/v3/cloud-client/requirements-test.txt
|
process
|
remove gcp devrel py tools from monitoring api cloud client requirements test txt remove gcp devrel py tools from monitoring api cloud client requirements test txt
| 1
|
21,572
| 29,924,949,792
|
IssuesEvent
|
2023-06-22 04:11:14
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
asyncssh 2.13.2 has 1 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/asyncssh
https://inspector.pypi.io/project/asyncssh
```{
"dependency": "asyncssh",
"version": "2.13.2",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "asyncssh-2.13.2/asyncssh/config.py:47",
"code": " return subprocess.run(cmd, check=False, shell=True, stdin=DEVNULL,\n stdout=DEVNULL, stderr=DEVNULL).returncode == 0",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpyavudvof/asyncssh"
}
}```
|
1.0
|
asyncssh 2.13.2 has 1 GuardDog issues - https://pypi.org/project/asyncssh
https://inspector.pypi.io/project/asyncssh
```{
"dependency": "asyncssh",
"version": "2.13.2",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "asyncssh-2.13.2/asyncssh/config.py:47",
"code": " return subprocess.run(cmd, check=False, shell=True, stdin=DEVNULL,\n stdout=DEVNULL, stderr=DEVNULL).returncode == 0",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpyavudvof/asyncssh"
}
}```
|
process
|
asyncssh has guarddog issues dependency asyncssh version result issues errors results silent process execution location asyncssh asyncssh config py code return subprocess run cmd check false shell true stdin devnull n stdout devnull stderr devnull returncode message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpyavudvof asyncssh
| 1
|
41,878
| 5,396,871,007
|
IssuesEvent
|
2017-02-27 13:08:24
|
hacken-in/hacken-in
|
https://api.github.com/repos/hacken-in/hacken-in
|
closed
|
Switch to Mapbox or improve map rendering
|
design reviewable
|
At the moment we're using [Leaflet](http://leafletjs.com/index.html) for rendering the maps – which is fine in terms of functionality – but the Maps look awful.
If we can improve the map rendering and stay with the OSM cards, I'm all for it.
Otherwise we could give Mapbox a try – I talked to @freenerd at yesterdays Devops meetup and I think they can arrange something if we bust the free plan of 3000 map views / month.
|
1.0
|
Switch to Mapbox or improve map rendering - At the moment we're using [Leaflet](http://leafletjs.com/index.html) for rendering the maps – which is fine in terms of functionality – but the Maps look awful.
If we can improve the map rendering and stay with the OSM cards, I'm all for it.
Otherwise we could give Mapbox a try – I talked to @freenerd at yesterdays Devops meetup and I think they can arrange something if we bust the free plan of 3000 map views / month.
|
non_process
|
switch to mapbox or improve map rendering at the moment we re using for rendering the maps – which is fine in terms of functionality – but the maps look awful if we can improve the map rendering and stay with the osm cards i m all for it otherwise we could give mapbox a try – i talked to freenerd at yesterdays devops meetup and i think they can arrange something if we bust the free plan of map views month
| 0
|
14,859
| 18,264,603,311
|
IssuesEvent
|
2021-10-04 06:48:15
|
Blazebit/blaze-persistence
|
https://api.github.com/repos/Blazebit/blaze-persistence
|
closed
|
blaze persistence annotation processor yet incompatible with java 17 LTS
|
kind: bug worth: high component: entity-view-annotation-processor
|
When compiling with java 17 (liberca / bellsoft) we get this exception
```
org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':compileJava'.
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.lambda$executeIfValid$1(ExecuteActionsTaskExecuter.java:188)
at org.gradle.internal.Try$Failure.ifSuccessfulOrElse(Try.java:282)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:186)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:174)
at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:109)
at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46)
at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:51)
at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57)
at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56)
at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52)
at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:74)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:402)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:389)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:382)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:368)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
Caused by: java.lang.RuntimeException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.invocationHelper(JavacTaskImpl.java:168)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.doCall(JavacTaskImpl.java:100)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.call(JavacTaskImpl.java:94)
at org.gradle.internal.compiler.java.IncrementalCompileTask.call(IncrementalCompileTask.java:89)
at org.gradle.api.internal.tasks.compile.AnnotationProcessingCompileTask.call(AnnotationProcessingCompileTask.java:94)
at org.gradle.api.internal.tasks.compile.ResourceCleaningCompilationTask.call(ResourceCleaningCompilationTask.java:57)
at org.gradle.api.internal.tasks.compile.JdkJavaCompiler.execute(JdkJavaCompiler.java:54)
at org.gradle.api.internal.tasks.compile.JdkJavaCompiler.execute(JdkJavaCompiler.java:39)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.delegateAndHandleErrors(NormalizingJavaCompiler.java:97)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.execute(NormalizingJavaCompiler.java:51)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.execute(NormalizingJavaCompiler.java:37)
at org.gradle.api.internal.tasks.compile.AnnotationProcessorDiscoveringCompiler.execute(AnnotationProcessorDiscoveringCompiler.java:51)
at org.gradle.api.internal.tasks.compile.AnnotationProcessorDiscoveringCompiler.execute(AnnotationProcessorDiscoveringCompiler.java:37)
at org.gradle.api.internal.tasks.compile.ModuleApplicationNameWritingCompiler.execute(ModuleApplicationNameWritingCompiler.java:46)
at org.gradle.api.internal.tasks.compile.ModuleApplicationNameWritingCompiler.execute(ModuleApplicationNameWritingCompiler.java:36)
at org.gradle.jvm.toolchain.internal.DefaultToolchainJavaCompiler.execute(DefaultToolchainJavaCompiler.java:57)
at org.gradle.api.tasks.compile.JavaCompile.lambda$createToolchainCompiler$1(JavaCompile.java:232)
at org.gradle.api.internal.tasks.compile.CleaningJavaCompiler.execute(CleaningJavaCompiler.java:53)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalCompilerFactory.lambda$createRebuildAllCompiler$0(IncrementalCompilerFactory.java:52)
at org.gradle.api.internal.tasks.compile.incremental.SelectiveCompiler.execute(SelectiveCompiler.java:67)
at org.gradle.api.internal.tasks.compile.incremental.SelectiveCompiler.execute(SelectiveCompiler.java:41)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalResultStoringCompiler.execute(IncrementalResultStoringCompiler.java:65)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalResultStoringCompiler.execute(IncrementalResultStoringCompiler.java:51)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler$2.call(CompileJavaBuildOperationReportingCompiler.java:59)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler$2.call(CompileJavaBuildOperationReportingCompiler.java:51)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler.execute(CompileJavaBuildOperationReportingCompiler.java:51)
at org.gradle.api.tasks.compile.JavaCompile.performCompilation(JavaCompile.java:279)
at org.gradle.api.tasks.compile.JavaCompile.performIncrementalCompilation(JavaCompile.java:165)
at org.gradle.api.tasks.compile.JavaCompile.compile(JavaCompile.java:146)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:104)
at org.gradle.api.internal.project.taskfactory.IncrementalInputsTaskAction.doExecute(IncrementalInputsTaskAction.java:32)
at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:51)
at org.gradle.api.internal.project.taskfactory.AbstractIncrementalTaskAction.execute(AbstractIncrementalTaskAction.java:25)
at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:29)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$2.run(ExecuteActionsTaskExecuter.java:502)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:56)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$run$1(DefaultBuildOperationExecutor.java:74)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.runWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:45)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:74)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:487)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:470)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.access$300(ExecuteActionsTaskExecuter.java:106)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.executeWithPreviousOutputFiles(ExecuteActionsTaskExecuter.java:271)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.execute(ExecuteActionsTaskExecuter.java:249)
at org.gradle.internal.execution.steps.ExecuteStep.executeInternal(ExecuteStep.java:89)
at org.gradle.internal.execution.steps.ExecuteStep.access$000(ExecuteStep.java:40)
at org.gradle.internal.execution.steps.ExecuteStep$1.call(ExecuteStep.java:53)
at org.gradle.internal.execution.steps.ExecuteStep$1.call(ExecuteStep.java:50)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:50)
at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:40)
at org.gradle.internal.execution.steps.RemovePreviousOutputsStep.execute(RemovePreviousOutputsStep.java:68)
at org.gradle.internal.execution.steps.RemovePreviousOutputsStep.execute(RemovePreviousOutputsStep.java:38)
at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:50)
at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:36)
at org.gradle.internal.execution.steps.CancelExecutionStep.execute(CancelExecutionStep.java:41)
at org.gradle.internal.execution.steps.TimeoutStep.executeWithoutTimeout(TimeoutStep.java:74)
at org.gradle.internal.execution.steps.TimeoutStep.execute(TimeoutStep.java:55)
at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:51)
at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:29)
at org.gradle.internal.execution.steps.CaptureStateAfterExecutionStep.execute(CaptureStateAfterExecutionStep.java:58)
at org.gradle.internal.execution.steps.CaptureStateAfterExecutionStep.execute(CaptureStateAfterExecutionStep.java:39)
at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:60)
at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:27)
at org.gradle.internal.execution.steps.BuildCacheStep.executeWithoutCache(BuildCacheStep.java:180)
at org.gradle.internal.execution.steps.BuildCacheStep.execute(BuildCacheStep.java:75)
at org.gradle.internal.execution.steps.BuildCacheStep.execute(BuildCacheStep.java:46)
at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:40)
at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:29)
at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:36)
at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:22)
at org.gradle.internal.execution.steps.SkipUpToDateStep.executeBecause(SkipUpToDateStep.java:105)
at org.gradle.internal.execution.steps.SkipUpToDateStep.lambda$execute$0(SkipUpToDateStep.java:98)
at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:53)
at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:37)
at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:85)
at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:42)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:37)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:27)
at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:92)
at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:50)
at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:114)
at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:57)
at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:73)
at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:47)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.lambda$execute$2(SkipEmptyWorkStep.java:92)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:92)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:33)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsStartedStep.execute(MarkSnapshottingInputsStartedStep.java:38)
at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:43)
at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:31)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.lambda$execute$0(AssignWorkspaceStep.java:40)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution$2.withWorkspace(ExecuteActionsTaskExecuter.java:284)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.execute(AssignWorkspaceStep.java:40)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.execute(AssignWorkspaceStep.java:30)
at org.gradle.internal.execution.steps.IdentityCacheStep.execute(IdentityCacheStep.java:37)
at org.gradle.internal.execution.steps.IdentityCacheStep.execute(IdentityCacheStep.java:27)
at org.gradle.internal.execution.steps.IdentifyStep.execute(IdentifyStep.java:44)
at org.gradle.internal.execution.steps.IdentifyStep.execute(IdentifyStep.java:33)
at org.gradle.internal.execution.impl.DefaultExecutionEngine$1.execute(DefaultExecutionEngine.java:76)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:185)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:174)
at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:109)
at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46)
at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:51)
at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57)
at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56)
at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52)
at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:74)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:402)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:389)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:382)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:368)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.await(EntityViewAnnotationProcessor.java:121)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.execute(EntityViewAnnotationProcessor.java:110)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.process(EntityViewAnnotationProcessor.java:97)
at org.gradle.api.internal.tasks.compile.processing.DelegatingProcessor.process(DelegatingProcessor.java:62)
at org.gradle.api.internal.tasks.compile.processing.IsolatingProcessor.process(IsolatingProcessor.java:50)
at org.gradle.api.internal.tasks.compile.processing.DelegatingProcessor.process(DelegatingProcessor.java:62)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.access$401(TimeTrackingProcessor.java:37)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor$5.create(TimeTrackingProcessor.java:99)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor$5.create(TimeTrackingProcessor.java:96)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.track(TimeTrackingProcessor.java:117)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.process(TimeTrackingProcessor.java:96)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.callProcessor(JavacProcessingEnvironment.java:1023)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.discoverAndRunProcs(JavacProcessingEnvironment.java:939)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment$Round.run(JavacProcessingEnvironment.java:1267)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.doProcessing(JavacProcessingEnvironment.java:1382)
at jdk.compiler/com.sun.tools.javac.main.JavaCompiler.processAnnotations(JavaCompiler.java:1234)
at jdk.compiler/com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:916)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.lambda$doCall$0(JavacTaskImpl.java:104)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.invocationHelper(JavacTaskImpl.java:152)
... 161 more
Caused by: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.await(EntityViewAnnotationProcessor.java:119)
... 179 more
Caused by: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at jdk.compiler/com.sun.tools.javac.model.JavacTypes.asMemberOf(JavacTypes.java:299)
at com.blazebit.persistence.view.processor.AttributeFilter.<init>(AttributeFilter.java:36)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaAttribute.addAttributeFilter(AnnotationMetaAttribute.java:365)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaAttribute.<init>(AnnotationMetaAttribute.java:203)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaSingularAttribute.<init>(AnnotationMetaSingularAttribute.java:31)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitDeclared(MetaAttributeGenerationVisitor.java:140)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitDeclared(MetaAttributeGenerationVisitor.java:44)
at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:1169)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitExecutable(MetaAttributeGenerationVisitor.java:157)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitExecutable(MetaAttributeGenerationVisitor.java:44)
at jdk.compiler/com.sun.tools.javac.code.Type$MethodType.accept(Type.java:1535)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaEntityView.<init>(AnnotationMetaEntityView.java:195)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor$7.run(EntityViewAnnotationProcessor.java:236)
```
Version: 1.16.2
JPA-Provider:
DBMS:
Application Server:
|
1.0
|
blaze persistence annotation processor yet incompatible with java 17 LTS - When compiling with java 17 (liberca / bellsoft) we get this exception
```
org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':compileJava'.
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.lambda$executeIfValid$1(ExecuteActionsTaskExecuter.java:188)
at org.gradle.internal.Try$Failure.ifSuccessfulOrElse(Try.java:282)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:186)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:174)
at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:109)
at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46)
at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:51)
at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57)
at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56)
at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52)
at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:74)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:402)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:389)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:382)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:368)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
Caused by: java.lang.RuntimeException: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.invocationHelper(JavacTaskImpl.java:168)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.doCall(JavacTaskImpl.java:100)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.call(JavacTaskImpl.java:94)
at org.gradle.internal.compiler.java.IncrementalCompileTask.call(IncrementalCompileTask.java:89)
at org.gradle.api.internal.tasks.compile.AnnotationProcessingCompileTask.call(AnnotationProcessingCompileTask.java:94)
at org.gradle.api.internal.tasks.compile.ResourceCleaningCompilationTask.call(ResourceCleaningCompilationTask.java:57)
at org.gradle.api.internal.tasks.compile.JdkJavaCompiler.execute(JdkJavaCompiler.java:54)
at org.gradle.api.internal.tasks.compile.JdkJavaCompiler.execute(JdkJavaCompiler.java:39)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.delegateAndHandleErrors(NormalizingJavaCompiler.java:97)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.execute(NormalizingJavaCompiler.java:51)
at org.gradle.api.internal.tasks.compile.NormalizingJavaCompiler.execute(NormalizingJavaCompiler.java:37)
at org.gradle.api.internal.tasks.compile.AnnotationProcessorDiscoveringCompiler.execute(AnnotationProcessorDiscoveringCompiler.java:51)
at org.gradle.api.internal.tasks.compile.AnnotationProcessorDiscoveringCompiler.execute(AnnotationProcessorDiscoveringCompiler.java:37)
at org.gradle.api.internal.tasks.compile.ModuleApplicationNameWritingCompiler.execute(ModuleApplicationNameWritingCompiler.java:46)
at org.gradle.api.internal.tasks.compile.ModuleApplicationNameWritingCompiler.execute(ModuleApplicationNameWritingCompiler.java:36)
at org.gradle.jvm.toolchain.internal.DefaultToolchainJavaCompiler.execute(DefaultToolchainJavaCompiler.java:57)
at org.gradle.api.tasks.compile.JavaCompile.lambda$createToolchainCompiler$1(JavaCompile.java:232)
at org.gradle.api.internal.tasks.compile.CleaningJavaCompiler.execute(CleaningJavaCompiler.java:53)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalCompilerFactory.lambda$createRebuildAllCompiler$0(IncrementalCompilerFactory.java:52)
at org.gradle.api.internal.tasks.compile.incremental.SelectiveCompiler.execute(SelectiveCompiler.java:67)
at org.gradle.api.internal.tasks.compile.incremental.SelectiveCompiler.execute(SelectiveCompiler.java:41)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalResultStoringCompiler.execute(IncrementalResultStoringCompiler.java:65)
at org.gradle.api.internal.tasks.compile.incremental.IncrementalResultStoringCompiler.execute(IncrementalResultStoringCompiler.java:51)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler$2.call(CompileJavaBuildOperationReportingCompiler.java:59)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler$2.call(CompileJavaBuildOperationReportingCompiler.java:51)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.compile.CompileJavaBuildOperationReportingCompiler.execute(CompileJavaBuildOperationReportingCompiler.java:51)
at org.gradle.api.tasks.compile.JavaCompile.performCompilation(JavaCompile.java:279)
at org.gradle.api.tasks.compile.JavaCompile.performIncrementalCompilation(JavaCompile.java:165)
at org.gradle.api.tasks.compile.JavaCompile.compile(JavaCompile.java:146)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:104)
at org.gradle.api.internal.project.taskfactory.IncrementalInputsTaskAction.doExecute(IncrementalInputsTaskAction.java:32)
at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:51)
at org.gradle.api.internal.project.taskfactory.AbstractIncrementalTaskAction.execute(AbstractIncrementalTaskAction.java:25)
at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:29)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$2.run(ExecuteActionsTaskExecuter.java:502)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:56)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$run$1(DefaultBuildOperationExecutor.java:74)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.runWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:45)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:74)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:487)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:470)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.access$300(ExecuteActionsTaskExecuter.java:106)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.executeWithPreviousOutputFiles(ExecuteActionsTaskExecuter.java:271)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.execute(ExecuteActionsTaskExecuter.java:249)
at org.gradle.internal.execution.steps.ExecuteStep.executeInternal(ExecuteStep.java:89)
at org.gradle.internal.execution.steps.ExecuteStep.access$000(ExecuteStep.java:40)
at org.gradle.internal.execution.steps.ExecuteStep$1.call(ExecuteStep.java:53)
at org.gradle.internal.execution.steps.ExecuteStep$1.call(ExecuteStep.java:50)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:50)
at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:40)
at org.gradle.internal.execution.steps.RemovePreviousOutputsStep.execute(RemovePreviousOutputsStep.java:68)
at org.gradle.internal.execution.steps.RemovePreviousOutputsStep.execute(RemovePreviousOutputsStep.java:38)
at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:50)
at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:36)
at org.gradle.internal.execution.steps.CancelExecutionStep.execute(CancelExecutionStep.java:41)
at org.gradle.internal.execution.steps.TimeoutStep.executeWithoutTimeout(TimeoutStep.java:74)
at org.gradle.internal.execution.steps.TimeoutStep.execute(TimeoutStep.java:55)
at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:51)
at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:29)
at org.gradle.internal.execution.steps.CaptureStateAfterExecutionStep.execute(CaptureStateAfterExecutionStep.java:58)
at org.gradle.internal.execution.steps.CaptureStateAfterExecutionStep.execute(CaptureStateAfterExecutionStep.java:39)
at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:60)
at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:27)
at org.gradle.internal.execution.steps.BuildCacheStep.executeWithoutCache(BuildCacheStep.java:180)
at org.gradle.internal.execution.steps.BuildCacheStep.execute(BuildCacheStep.java:75)
at org.gradle.internal.execution.steps.BuildCacheStep.execute(BuildCacheStep.java:46)
at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:40)
at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:29)
at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:36)
at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:22)
at org.gradle.internal.execution.steps.SkipUpToDateStep.executeBecause(SkipUpToDateStep.java:105)
at org.gradle.internal.execution.steps.SkipUpToDateStep.lambda$execute$0(SkipUpToDateStep.java:98)
at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:53)
at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:37)
at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:85)
at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:42)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:37)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:27)
at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:92)
at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:50)
at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:114)
at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:57)
at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:73)
at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:47)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.lambda$execute$2(SkipEmptyWorkStep.java:92)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:92)
at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:33)
at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsStartedStep.execute(MarkSnapshottingInputsStartedStep.java:38)
at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:43)
at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:31)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.lambda$execute$0(AssignWorkspaceStep.java:40)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution$2.withWorkspace(ExecuteActionsTaskExecuter.java:284)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.execute(AssignWorkspaceStep.java:40)
at org.gradle.internal.execution.steps.AssignWorkspaceStep.execute(AssignWorkspaceStep.java:30)
at org.gradle.internal.execution.steps.IdentityCacheStep.execute(IdentityCacheStep.java:37)
at org.gradle.internal.execution.steps.IdentityCacheStep.execute(IdentityCacheStep.java:27)
at org.gradle.internal.execution.steps.IdentifyStep.execute(IdentifyStep.java:44)
at org.gradle.internal.execution.steps.IdentifyStep.execute(IdentifyStep.java:33)
at org.gradle.internal.execution.impl.DefaultExecutionEngine$1.execute(DefaultExecutionEngine.java:76)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:185)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:174)
at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:109)
at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46)
at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:51)
at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57)
at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56)
at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:200)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:195)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:75)
at org.gradle.internal.operations.DefaultBuildOperationRunner$3.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:153)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:68)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:62)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.lambda$call$2(DefaultBuildOperationExecutor.java:79)
at org.gradle.internal.operations.UnmanagedBuildOperationWrapper.callWithUnmanagedSupport(UnmanagedBuildOperationWrapper.java:54)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:79)
at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52)
at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:74)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:402)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:389)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:382)
at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:368)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182)
at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124)
at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.await(EntityViewAnnotationProcessor.java:121)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.execute(EntityViewAnnotationProcessor.java:110)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.process(EntityViewAnnotationProcessor.java:97)
at org.gradle.api.internal.tasks.compile.processing.DelegatingProcessor.process(DelegatingProcessor.java:62)
at org.gradle.api.internal.tasks.compile.processing.IsolatingProcessor.process(IsolatingProcessor.java:50)
at org.gradle.api.internal.tasks.compile.processing.DelegatingProcessor.process(DelegatingProcessor.java:62)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.access$401(TimeTrackingProcessor.java:37)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor$5.create(TimeTrackingProcessor.java:99)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor$5.create(TimeTrackingProcessor.java:96)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.track(TimeTrackingProcessor.java:117)
at org.gradle.api.internal.tasks.compile.processing.TimeTrackingProcessor.process(TimeTrackingProcessor.java:96)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.callProcessor(JavacProcessingEnvironment.java:1023)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.discoverAndRunProcs(JavacProcessingEnvironment.java:939)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment$Round.run(JavacProcessingEnvironment.java:1267)
at jdk.compiler/com.sun.tools.javac.processing.JavacProcessingEnvironment.doProcessing(JavacProcessingEnvironment.java:1382)
at jdk.compiler/com.sun.tools.javac.main.JavaCompiler.processAnnotations(JavaCompiler.java:1234)
at jdk.compiler/com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:916)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.lambda$doCall$0(JavacTaskImpl.java:104)
at jdk.compiler/com.sun.tools.javac.api.JavacTaskImpl.invocationHelper(JavacTaskImpl.java:152)
... 161 more
Caused by: java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor.await(EntityViewAnnotationProcessor.java:119)
... 179 more
Caused by: java.lang.IllegalArgumentException: FilterValue@com.blazebit.persistence.view.filter.EqualFilter<FilterValue>
at jdk.compiler/com.sun.tools.javac.model.JavacTypes.asMemberOf(JavacTypes.java:299)
at com.blazebit.persistence.view.processor.AttributeFilter.<init>(AttributeFilter.java:36)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaAttribute.addAttributeFilter(AnnotationMetaAttribute.java:365)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaAttribute.<init>(AnnotationMetaAttribute.java:203)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaSingularAttribute.<init>(AnnotationMetaSingularAttribute.java:31)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitDeclared(MetaAttributeGenerationVisitor.java:140)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitDeclared(MetaAttributeGenerationVisitor.java:44)
at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:1169)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitExecutable(MetaAttributeGenerationVisitor.java:157)
at com.blazebit.persistence.view.processor.annotation.MetaAttributeGenerationVisitor.visitExecutable(MetaAttributeGenerationVisitor.java:44)
at jdk.compiler/com.sun.tools.javac.code.Type$MethodType.accept(Type.java:1535)
at com.blazebit.persistence.view.processor.annotation.AnnotationMetaEntityView.<init>(AnnotationMetaEntityView.java:195)
at com.blazebit.persistence.view.processor.EntityViewAnnotationProcessor$7.run(EntityViewAnnotationProcessor.java:236)
```
Version: 1.16.2
JPA-Provider:
DBMS:
Application Server:
|
process
|
blaze persistence annotation processor yet incompatible with java lts when compiling with java liberca bellsoft we get this exception org gradle api tasks taskexecutionexception execution failed for task compilejava at org gradle api internal tasks execution executeactionstaskexecuter lambda executeifvalid executeactionstaskexecuter java at org gradle internal try failure ifsuccessfulorelse try java at org gradle api internal tasks execution executeactionstaskexecuter executeifvalid executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter execute executeactionstaskexecuter java at org gradle api internal tasks execution cleanupstaleoutputsexecuter execute cleanupstaleoutputsexecuter java at org gradle api internal tasks execution finalizepropertiestaskexecuter execute finalizepropertiestaskexecuter java at org gradle api internal tasks execution resolvetaskexecutionmodeexecuter execute resolvetaskexecutionmodeexecuter java at org gradle api internal tasks execution skiptaskwithnoactionsexecuter execute skiptaskwithnoactionsexecuter java at org gradle api internal tasks execution skiponlyiftaskexecuter execute skiponlyiftaskexecuter java at org gradle api internal tasks execution catchexceptiontaskexecuter execute catchexceptiontaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter executetask eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor lambda call defaultbuildoperationexecutor java at org gradle internal operations unmanagedbuildoperationwrapper callwithunmanagedsupport unmanagedbuildoperationwrapper java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle api internal tasks execution eventfiringtaskexecuter execute eventfiringtaskexecuter java at org gradle execution plan localtasknodeexecutor execute localtasknodeexecutor java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution plan defaultplanexecutor executorworker lambda run defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker execute defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker executenextnode defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker run defaultplanexecutor java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java caused by java lang runtimeexception java lang runtimeexception java util concurrent executionexception java lang illegalargumentexception filtervalue com blazebit persistence view filter equalfilter at jdk compiler com sun tools javac api javactaskimpl invocationhelper javactaskimpl java at jdk compiler com sun tools javac api javactaskimpl docall javactaskimpl java at jdk compiler com sun tools javac api javactaskimpl call javactaskimpl java at org gradle internal compiler java incrementalcompiletask call incrementalcompiletask java at org gradle api internal tasks compile annotationprocessingcompiletask call annotationprocessingcompiletask java at org gradle api internal tasks compile resourcecleaningcompilationtask call resourcecleaningcompilationtask java at org gradle api internal tasks compile jdkjavacompiler execute jdkjavacompiler java at org gradle api internal tasks compile jdkjavacompiler execute jdkjavacompiler java at org gradle api internal tasks compile normalizingjavacompiler delegateandhandleerrors normalizingjavacompiler java at org gradle api internal tasks compile normalizingjavacompiler execute normalizingjavacompiler java at org gradle api internal tasks compile normalizingjavacompiler execute normalizingjavacompiler java at org gradle api internal tasks compile annotationprocessordiscoveringcompiler execute annotationprocessordiscoveringcompiler java at org gradle api internal tasks compile annotationprocessordiscoveringcompiler execute annotationprocessordiscoveringcompiler java at org gradle api internal tasks compile moduleapplicationnamewritingcompiler execute moduleapplicationnamewritingcompiler java at org gradle api internal tasks compile moduleapplicationnamewritingcompiler execute moduleapplicationnamewritingcompiler java at org gradle jvm toolchain internal defaulttoolchainjavacompiler execute defaulttoolchainjavacompiler java at org gradle api tasks compile javacompile lambda createtoolchaincompiler javacompile java at org gradle api internal tasks compile cleaningjavacompiler execute cleaningjavacompiler java at org gradle api internal tasks compile incremental incrementalcompilerfactory lambda createrebuildallcompiler incrementalcompilerfactory java at org gradle api internal tasks compile incremental selectivecompiler execute selectivecompiler java at org gradle api internal tasks compile incremental selectivecompiler execute selectivecompiler java at org gradle api internal tasks compile incremental incrementalresultstoringcompiler execute incrementalresultstoringcompiler java at org gradle api internal tasks compile incremental incrementalresultstoringcompiler execute incrementalresultstoringcompiler java at org gradle api internal tasks compile compilejavabuildoperationreportingcompiler call compilejavabuildoperationreportingcompiler java at org gradle api internal tasks compile compilejavabuildoperationreportingcompiler call compilejavabuildoperationreportingcompiler java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor lambda call defaultbuildoperationexecutor java at org gradle internal operations unmanagedbuildoperationwrapper callwithunmanagedsupport unmanagedbuildoperationwrapper java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle api internal tasks compile compilejavabuildoperationreportingcompiler execute compilejavabuildoperationreportingcompiler java at org gradle api tasks compile javacompile performcompilation javacompile java at org gradle api tasks compile javacompile performincrementalcompilation javacompile java at org gradle api tasks compile javacompile compile javacompile java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at org gradle internal reflect javamethod invoke javamethod java at org gradle api internal project taskfactory incrementalinputstaskaction doexecute incrementalinputstaskaction java at org gradle api internal project taskfactory standardtaskaction execute standardtaskaction java at org gradle api internal project taskfactory abstractincrementaltaskaction execute abstractincrementaltaskaction java at org gradle api internal project taskfactory standardtaskaction execute standardtaskaction java at org gradle api internal tasks execution executeactionstaskexecuter run executeactionstaskexecuter java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor lambda run defaultbuildoperationexecutor java at org gradle internal operations unmanagedbuildoperationwrapper runwithunmanagedsupport unmanagedbuildoperationwrapper java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal tasks execution executeactionstaskexecuter executeaction executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter executeactions executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter access executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter taskexecution executewithpreviousoutputfiles executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter taskexecution execute executeactionstaskexecuter java at org gradle internal execution steps executestep executeinternal executestep java at org gradle internal execution steps executestep access executestep java at org gradle internal execution steps executestep call executestep java at org gradle internal execution steps executestep call executestep java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor lambda call defaultbuildoperationexecutor java at org gradle internal operations unmanagedbuildoperationwrapper callwithunmanagedsupport unmanagedbuildoperationwrapper java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle internal execution steps executestep execute executestep java at org gradle internal execution steps executestep execute executestep java at org gradle internal execution steps removepreviousoutputsstep execute removepreviousoutputsstep java at org gradle internal execution steps removepreviousoutputsstep execute removepreviousoutputsstep java at org gradle internal execution steps resolveinputchangesstep execute resolveinputchangesstep java at org gradle internal execution steps resolveinputchangesstep execute resolveinputchangesstep java at org gradle internal execution steps cancelexecutionstep execute cancelexecutionstep java at org gradle internal execution steps timeoutstep executewithouttimeout timeoutstep java at org gradle internal execution steps timeoutstep execute timeoutstep java at org gradle internal execution steps createoutputsstep execute createoutputsstep java at org gradle internal execution steps createoutputsstep execute createoutputsstep java at org gradle internal execution steps capturestateafterexecutionstep execute capturestateafterexecutionstep java at org gradle internal execution steps capturestateafterexecutionstep execute capturestateafterexecutionstep java at org gradle internal execution steps broadcastchangingoutputsstep execute broadcastchangingoutputsstep java at org gradle internal execution steps broadcastchangingoutputsstep execute broadcastchangingoutputsstep java at org gradle internal execution steps buildcachestep executewithoutcache buildcachestep java at org gradle internal execution steps buildcachestep execute buildcachestep java at org gradle internal execution steps buildcachestep execute buildcachestep java at org gradle internal execution steps storeexecutionstatestep execute storeexecutionstatestep java at org gradle internal execution steps storeexecutionstatestep execute storeexecutionstatestep java at org gradle internal execution steps recordoutputsstep execute recordoutputsstep java at org gradle internal execution steps recordoutputsstep execute recordoutputsstep java at org gradle internal execution steps skipuptodatestep executebecause skipuptodatestep java at org gradle internal execution steps skipuptodatestep lambda execute skipuptodatestep java at org gradle internal execution steps skipuptodatestep execute skipuptodatestep java at org gradle internal execution steps skipuptodatestep execute skipuptodatestep java at org gradle internal execution steps resolvechangesstep execute resolvechangesstep java at org gradle internal execution steps resolvechangesstep execute resolvechangesstep java at org gradle internal execution steps legacy marksnapshottinginputsfinishedstep execute marksnapshottinginputsfinishedstep java at org gradle internal execution steps legacy marksnapshottinginputsfinishedstep execute marksnapshottinginputsfinishedstep java at org gradle internal execution steps resolvecachingstatestep execute resolvecachingstatestep java at org gradle internal execution steps resolvecachingstatestep execute resolvecachingstatestep java at org gradle internal execution steps validatestep execute validatestep java at org gradle internal execution steps validatestep execute validatestep java at org gradle internal execution steps capturestatebeforeexecutionstep execute capturestatebeforeexecutionstep java at org gradle internal execution steps capturestatebeforeexecutionstep execute capturestatebeforeexecutionstep java at org gradle internal execution steps skipemptyworkstep lambda execute skipemptyworkstep java at org gradle internal execution steps skipemptyworkstep execute skipemptyworkstep java at org gradle internal execution steps skipemptyworkstep execute skipemptyworkstep java at org gradle internal execution steps legacy marksnapshottinginputsstartedstep execute marksnapshottinginputsstartedstep java at org gradle internal execution steps loadexecutionstatestep execute loadexecutionstatestep java at org gradle internal execution steps loadexecutionstatestep execute loadexecutionstatestep java at org gradle internal execution steps assignworkspacestep lambda execute assignworkspacestep java at org gradle api internal tasks execution executeactionstaskexecuter taskexecution withworkspace executeactionstaskexecuter java at org gradle internal execution steps assignworkspacestep execute assignworkspacestep java at org gradle internal execution steps assignworkspacestep execute assignworkspacestep java at org gradle internal execution steps identitycachestep execute identitycachestep java at org gradle internal execution steps identitycachestep execute identitycachestep java at org gradle internal execution steps identifystep execute identifystep java at org gradle internal execution steps identifystep execute identifystep java at org gradle internal execution impl defaultexecutionengine execute defaultexecutionengine java at org gradle api internal tasks execution executeactionstaskexecuter executeifvalid executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter execute executeactionstaskexecuter java at org gradle api internal tasks execution cleanupstaleoutputsexecuter execute cleanupstaleoutputsexecuter java at org gradle api internal tasks execution finalizepropertiestaskexecuter execute finalizepropertiestaskexecuter java at org gradle api internal tasks execution resolvetaskexecutionmodeexecuter execute resolvetaskexecutionmodeexecuter java at org gradle api internal tasks execution skiptaskwithnoactionsexecuter execute skiptaskwithnoactionsexecuter java at org gradle api internal tasks execution skiponlyiftaskexecuter execute skiponlyiftaskexecuter java at org gradle api internal tasks execution catchexceptiontaskexecuter execute catchexceptiontaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter executetask eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor lambda call defaultbuildoperationexecutor java at org gradle internal operations unmanagedbuildoperationwrapper callwithunmanagedsupport unmanagedbuildoperationwrapper java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle api internal tasks execution eventfiringtaskexecuter execute eventfiringtaskexecuter java at org gradle execution plan localtasknodeexecutor execute localtasknodeexecutor java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution plan defaultplanexecutor executorworker lambda run defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker execute defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker executenextnode defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker run defaultplanexecutor java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java caused by java lang runtimeexception java util concurrent executionexception java lang illegalargumentexception filtervalue com blazebit persistence view filter equalfilter at com blazebit persistence view processor entityviewannotationprocessor await entityviewannotationprocessor java at com blazebit persistence view processor entityviewannotationprocessor execute entityviewannotationprocessor java at com blazebit persistence view processor entityviewannotationprocessor process entityviewannotationprocessor java at org gradle api internal tasks compile processing delegatingprocessor process delegatingprocessor java at org gradle api internal tasks compile processing isolatingprocessor process isolatingprocessor java at org gradle api internal tasks compile processing delegatingprocessor process delegatingprocessor java at org gradle api internal tasks compile processing timetrackingprocessor access timetrackingprocessor java at org gradle api internal tasks compile processing timetrackingprocessor create timetrackingprocessor java at org gradle api internal tasks compile processing timetrackingprocessor create timetrackingprocessor java at org gradle api internal tasks compile processing timetrackingprocessor track timetrackingprocessor java at org gradle api internal tasks compile processing timetrackingprocessor process timetrackingprocessor java at jdk compiler com sun tools javac processing javacprocessingenvironment callprocessor javacprocessingenvironment java at jdk compiler com sun tools javac processing javacprocessingenvironment discoverandrunprocs javacprocessingenvironment java at jdk compiler com sun tools javac processing javacprocessingenvironment round run javacprocessingenvironment java at jdk compiler com sun tools javac processing javacprocessingenvironment doprocessing javacprocessingenvironment java at jdk compiler com sun tools javac main javacompiler processannotations javacompiler java at jdk compiler com sun tools javac main javacompiler compile javacompiler java at jdk compiler com sun tools javac api javactaskimpl lambda docall javactaskimpl java at jdk compiler com sun tools javac api javactaskimpl invocationhelper javactaskimpl java more caused by java util concurrent executionexception java lang illegalargumentexception filtervalue com blazebit persistence view filter equalfilter at com blazebit persistence view processor entityviewannotationprocessor await entityviewannotationprocessor java more caused by java lang illegalargumentexception filtervalue com blazebit persistence view filter equalfilter at jdk compiler com sun tools javac model javactypes asmemberof javactypes java at com blazebit persistence view processor attributefilter attributefilter java at com blazebit persistence view processor annotation annotationmetaattribute addattributefilter annotationmetaattribute java at com blazebit persistence view processor annotation annotationmetaattribute annotationmetaattribute java at com blazebit persistence view processor annotation annotationmetasingularattribute annotationmetasingularattribute java at com blazebit persistence view processor annotation metaattributegenerationvisitor visitdeclared metaattributegenerationvisitor java at com blazebit persistence view processor annotation metaattributegenerationvisitor visitdeclared metaattributegenerationvisitor java at jdk compiler com sun tools javac code type classtype accept type java at com blazebit persistence view processor annotation metaattributegenerationvisitor visitexecutable metaattributegenerationvisitor java at com blazebit persistence view processor annotation metaattributegenerationvisitor visitexecutable metaattributegenerationvisitor java at jdk compiler com sun tools javac code type methodtype accept type java at com blazebit persistence view processor annotation annotationmetaentityview annotationmetaentityview java at com blazebit persistence view processor entityviewannotationprocessor run entityviewannotationprocessor java version jpa provider dbms application server
| 1
|
266,269
| 20,143,826,264
|
IssuesEvent
|
2022-02-09 04:03:22
|
gromeron/git_web_practice
|
https://api.github.com/repos/gromeron/git_web_practice
|
closed
|
Un commit que no sigue la convención de código o FIX a realizar
|
documentation
|
La convención del mensaje del último commit no es la esperada:
`FIX1 - correcciones pagina5 y pagina3`
Recuerde que debe tener el siguiente formato: `<Identificador de la corrección>: <Comentario>`
Para realizar la corrección del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f`
Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
|
1.0
|
Un commit que no sigue la convención de código o FIX a realizar - La convención del mensaje del último commit no es la esperada:
`FIX1 - correcciones pagina5 y pagina3`
Recuerde que debe tener el siguiente formato: `<Identificador de la corrección>: <Comentario>`
Para realizar la corrección del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f`
Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
|
non_process
|
un commit que no sigue la convención de código o fix a realizar la convención del mensaje del último commit no es la esperada correcciones y recuerde que debe tener el siguiente formato para realizar la corrección del mensaje de commit ejecute los comandos git commit amend y git push f este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado
| 0
|
71,570
| 9,528,536,592
|
IssuesEvent
|
2019-04-29 08:44:04
|
fergiemcdowall/search-index
|
https://api.github.com/repos/fergiemcdowall/search-index
|
opened
|
How to use the different parts of the API?
|
documentation question
|
Is it possible to do stuff like:
* autocomplete on several words (matching on two, three or more words, meaning this needs to be indexed too)
* number of results
* only show a subset (i.e. 10 of 273)
* filter on more than just one word
* delete everything in the index
I was thinking if I could get some hint on what is possible now and in what direction I need to check, I could do some testing myself. And make it easily available in the `[norch-vue](https://github.com/eklem/norch-vue)`
|
1.0
|
How to use the different parts of the API? - Is it possible to do stuff like:
* autocomplete on several words (matching on two, three or more words, meaning this needs to be indexed too)
* number of results
* only show a subset (i.e. 10 of 273)
* filter on more than just one word
* delete everything in the index
I was thinking if I could get some hint on what is possible now and in what direction I need to check, I could do some testing myself. And make it easily available in the `[norch-vue](https://github.com/eklem/norch-vue)`
|
non_process
|
how to use the different parts of the api is it possible to do stuff like autocomplete on several words matching on two three or more words meaning this needs to be indexed too number of results only show a subset i e of filter on more than just one word delete everything in the index i was thinking if i could get some hint on what is possible now and in what direction i need to check i could do some testing myself and make it easily available in the
| 0
|
317,931
| 27,273,611,046
|
IssuesEvent
|
2023-02-23 01:40:11
|
powbot/issues
|
https://api.github.com/repos/powbot/issues
|
closed
|
[BUG] Npc health percent returns 1.0 when an npc has 0.0% and is dying
|
bug needs-test
|
**Describe the bug**
An npc returns 0.0% briefly when you kill them, but when their dying animation begins the health resets to 1.0.
**To Reproduce/Line of code causing issue**
Steps to reproduce the behavior:
1. Add npc health percent and animation to paint builder
2. Kill the npc
3. Watch the health percent reset to 1.0 when their dying animation begins
**Expected behavior**
Npc health should stay at 0.0 while they/their health bar remains present
|
1.0
|
[BUG] Npc health percent returns 1.0 when an npc has 0.0% and is dying - **Describe the bug**
An npc returns 0.0% briefly when you kill them, but when their dying animation begins the health resets to 1.0.
**To Reproduce/Line of code causing issue**
Steps to reproduce the behavior:
1. Add npc health percent and animation to paint builder
2. Kill the npc
3. Watch the health percent reset to 1.0 when their dying animation begins
**Expected behavior**
Npc health should stay at 0.0 while they/their health bar remains present
|
non_process
|
npc health percent returns when an npc has and is dying describe the bug an npc returns briefly when you kill them but when their dying animation begins the health resets to to reproduce line of code causing issue steps to reproduce the behavior add npc health percent and animation to paint builder kill the npc watch the health percent reset to when their dying animation begins expected behavior npc health should stay at while they their health bar remains present
| 0
|
276,794
| 30,543,961,723
|
IssuesEvent
|
2023-07-20 01:10:54
|
Yoavmartin/vulnerable-node
|
https://api.github.com/repos/Yoavmartin/vulnerable-node
|
opened
|
CVE-2018-19826 (Medium) detected in node-sass3.5.5
|
Mend: dependency security vulnerability
|
## CVE-2018-19826 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass3.5.5</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/node-sass/src/libsass/src/inspect.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-19826>CVE-2018-19826</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
True
|
CVE-2018-19826 (Medium) detected in node-sass3.5.5 - ## CVE-2018-19826 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass3.5.5</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/node-sass/src/libsass/src/inspect.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-19826>CVE-2018-19826</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in node cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href found in base branch master vulnerable source files node modules node sass src libsass src inspect cpp vulnerability details disputed in inspect cpp in libsass a high memory footprint caused by an endless loop containing a sass inspect operator sass string quoted stack frame may cause a denial of service via crafted sass input files with stray or characters note upstream comments indicate this issue is closed as won t fix and works as intended by design publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
| 0
|
4,205
| 7,165,061,012
|
IssuesEvent
|
2018-01-29 13:21:21
|
ncbo/bioportal-project
|
https://api.github.com/repos/ncbo/bioportal-project
|
closed
|
ICD-O-3_1: pull URL points to an HTML page
|
ontology processing problem
|
We received a [complaint on the support list](http://ncbo-support.2288202.n4.nabble.com/bioontology-support-Ontology-Set-Up-td4656042.html) that the [ICD-O-3_1 ontology](http://bioportal.bioontology.org/ontologies/ICD-O-3_1) didn't parse. The pull URL points to an HTML page. It's unclear after some initial investigation that this ontology is available in any format we accept. Waiting for confirmation from user, but if we don't get any, the pull URL should be disabled (and perhaps the ontology entry deleted).
_Note that this is a private ontology._
|
1.0
|
ICD-O-3_1: pull URL points to an HTML page - We received a [complaint on the support list](http://ncbo-support.2288202.n4.nabble.com/bioontology-support-Ontology-Set-Up-td4656042.html) that the [ICD-O-3_1 ontology](http://bioportal.bioontology.org/ontologies/ICD-O-3_1) didn't parse. The pull URL points to an HTML page. It's unclear after some initial investigation that this ontology is available in any format we accept. Waiting for confirmation from user, but if we don't get any, the pull URL should be disabled (and perhaps the ontology entry deleted).
_Note that this is a private ontology._
|
process
|
icd o pull url points to an html page we received a that the didn t parse the pull url points to an html page it s unclear after some initial investigation that this ontology is available in any format we accept waiting for confirmation from user but if we don t get any the pull url should be disabled and perhaps the ontology entry deleted note that this is a private ontology
| 1
|
302
| 2,735,133,990
|
IssuesEvent
|
2015-04-18 03:39:27
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Activate fast-forward from mysql_users
|
ADMIN AUTHENTICATION CONNECTION POOL cxx_pa development MYSQL PROTOCOL QUERY PROCESSOR ROUTING
|
Add one more column to mysql_users table to specify if a specific user should have fast forwarding enabled
|
1.0
|
Activate fast-forward from mysql_users - Add one more column to mysql_users table to specify if a specific user should have fast forwarding enabled
|
process
|
activate fast forward from mysql users add one more column to mysql users table to specify if a specific user should have fast forwarding enabled
| 1
|
18,099
| 24,125,382,519
|
IssuesEvent
|
2022-09-20 23:28:15
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
No security-lib-ci yaml for reference
|
devops/prod Pri1 devops-cicd-process/tech
|
[Enter feedback here]
Where is the security-lib-ci yaml?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
1.0
|
No security-lib-ci yaml for reference -
[Enter feedback here]
Where is the security-lib-ci yaml?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
process
|
no security lib ci yaml for reference where is the security lib ci yaml document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
| 1
|
125,301
| 12,256,630,436
|
IssuesEvent
|
2020-05-06 12:26:15
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Missing good enough C++ platform portability and cross compilation instructions
|
area/documentation kind/question lang/other priority/P3
|
<!--
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers here:
- grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
- StackOverflow, with "grpc" tag: http://stackoverflow.com/questions/tagged/grpc
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
Latest version
C++
### What operating system (Linux, Windows, …) and version?
Build = Windows 10 x64
Target = Segger embOS on Cortex-M3/M4 based MCUs
### What runtime / compiler are you using (e.g. python version or version of gcc)
IAR ARM
### What did you do?
If possible, provide a recipe for reproducing the error. Try being specific and include code snippets if helpful.
1. Downloaded latest release from [releases](https://github.com/grpc/grpc/releases).
2. Tried to compile that code in IAR ARM, coming to conclusion that cross compilation of gRPC is required to run gRPC on a truly aka single chip aka bare metal embedded system.
3. Tried to cross compile according to [issue 9719](https://github.com/grpc/grpc/issues/9719), coming to conclusion that cross compilation only seems possible with gcc, since several language specifics like __inline, __sync_synchronize, __volatile__,... are being used.
**=> Stuck.**
### What did you expect to see?
Good enough instructions on how to cross compile to an truly embedded target. After all, https://grpc.io/blog/installation says "easy installation" and https://grpc.io/blog/principles "viable on CPU & memory limited devices".
### What did you see instead?
Several issues describing issues with cross compilation.
No known use of gRPC on an ARM based truly embedded MCU.
<!--
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md for how to diagnose problems better.
-->
### Anything else we should know about your project / environment?
On truly embedded systems, IAR ARM is among the four mostly used compilers (ARM, GCC, IAR, Keil).
|
1.0
|
Missing good enough C++ platform portability and cross compilation instructions - <!--
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers here:
- grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
- StackOverflow, with "grpc" tag: http://stackoverflow.com/questions/tagged/grpc
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
Latest version
C++
### What operating system (Linux, Windows, …) and version?
Build = Windows 10 x64
Target = Segger embOS on Cortex-M3/M4 based MCUs
### What runtime / compiler are you using (e.g. python version or version of gcc)
IAR ARM
### What did you do?
If possible, provide a recipe for reproducing the error. Try being specific and include code snippets if helpful.
1. Downloaded latest release from [releases](https://github.com/grpc/grpc/releases).
2. Tried to compile that code in IAR ARM, coming to conclusion that cross compilation of gRPC is required to run gRPC on a truly aka single chip aka bare metal embedded system.
3. Tried to cross compile according to [issue 9719](https://github.com/grpc/grpc/issues/9719), coming to conclusion that cross compilation only seems possible with gcc, since several language specifics like __inline, __sync_synchronize, __volatile__,... are being used.
**=> Stuck.**
### What did you expect to see?
Good enough instructions on how to cross compile to an truly embedded target. After all, https://grpc.io/blog/installation says "easy installation" and https://grpc.io/blog/principles "viable on CPU & memory limited devices".
### What did you see instead?
Several issues describing issues with cross compilation.
No known use of gRPC on an ARM based truly embedded MCU.
<!--
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md for how to diagnose problems better.
-->
### Anything else we should know about your project / environment?
On truly embedded systems, IAR ARM is among the four mostly used compilers (ARM, GCC, IAR, Keil).
|
non_process
|
missing good enough c platform portability and cross compilation instructions this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers here grpc io mailing list stackoverflow with grpc tag issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using latest version c what operating system linux windows … and version build windows target segger embos on cortex based mcus what runtime compiler are you using e g python version or version of gcc iar arm what did you do if possible provide a recipe for reproducing the error try being specific and include code snippets if helpful downloaded latest release from tried to compile that code in iar arm coming to conclusion that cross compilation of grpc is required to run grpc on a truly aka single chip aka bare metal embedded system tried to cross compile according to coming to conclusion that cross compilation only seems possible with gcc since several language specifics like inline sync synchronize volatile are being used stuck what did you expect to see good enough instructions on how to cross compile to an truly embedded target after all says easy installation and viable on cpu memory limited devices what did you see instead several issues describing issues with cross compilation no known use of grpc on an arm based truly embedded mcu make sure you include information that can help us debug full error message exception listing stack trace logs see for how to diagnose problems better anything else we should know about your project environment on truly embedded systems iar arm is among the four mostly used compilers arm gcc iar keil
| 0
|
20,204
| 26,782,778,351
|
IssuesEvent
|
2023-01-31 22:52:07
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Release X.Y.Z - $MONTH $YEAR
|
P1 type: process release team-OSS
|
# Status of Bazel X.Y.Z
<!-- The first item is only needed for major releases (X.0.0) -->
- Target baseline: [date]
- Expected release date: [date]
- [List of release blockers](link-to-milestone)
To report a release-blocking bug, please add a comment with the text `@bazel-io flag` to the issue. A release manager will triage it and add it to the milestone.
To cherry-pick a mainline commit into X.Y.Z, simply send a PR against the `release-X.Y.Z` branch.
**Task list:**
<!-- The first three items are only needed for major releases (X.0.0) -->
- [ ] Pick release baseline: [link to base commit]
- [ ] Create release candidate: X.Y.Zrc1
- [ ] Check downstream projects
- [ ] Create [draft release announcement](https://docs.google.com/document/d/1pu2ARPweOCTxPsRR8snoDtkC9R51XWRyBXeiC6Ql5so/edit) <!-- Note that there should be a new Bazel Release Announcement document for every major release. For minor and patch releases, use the latest open doc. -->
- [ ] Send the release announcement PR for review: [link to bazel-blog PR] <!-- Only for major releases. -->
- [ ] Push the release and notify package maintainers: [link to comment notifying package maintainers]
- [ ] Update the documentation
- [ ] Push the blog post: [link to blog post] <!-- Only for major releases. -->
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
1.0
|
Release X.Y.Z - $MONTH $YEAR - # Status of Bazel X.Y.Z
<!-- The first item is only needed for major releases (X.0.0) -->
- Target baseline: [date]
- Expected release date: [date]
- [List of release blockers](link-to-milestone)
To report a release-blocking bug, please add a comment with the text `@bazel-io flag` to the issue. A release manager will triage it and add it to the milestone.
To cherry-pick a mainline commit into X.Y.Z, simply send a PR against the `release-X.Y.Z` branch.
**Task list:**
<!-- The first three items are only needed for major releases (X.0.0) -->
- [ ] Pick release baseline: [link to base commit]
- [ ] Create release candidate: X.Y.Zrc1
- [ ] Check downstream projects
- [ ] Create [draft release announcement](https://docs.google.com/document/d/1pu2ARPweOCTxPsRR8snoDtkC9R51XWRyBXeiC6Ql5so/edit) <!-- Note that there should be a new Bazel Release Announcement document for every major release. For minor and patch releases, use the latest open doc. -->
- [ ] Send the release announcement PR for review: [link to bazel-blog PR] <!-- Only for major releases. -->
- [ ] Push the release and notify package maintainers: [link to comment notifying package maintainers]
- [ ] Update the documentation
- [ ] Push the blog post: [link to blog post] <!-- Only for major releases. -->
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
process
|
release x y z month year status of bazel x y z target baseline expected release date link to milestone to report a release blocking bug please add a comment with the text bazel io flag to the issue a release manager will triage it and add it to the milestone to cherry pick a mainline commit into x y z simply send a pr against the release x y z branch task list pick release baseline create release candidate x y check downstream projects create send the release announcement pr for review push the release and notify package maintainers update the documentation push the blog post update the
| 1
|
760,151
| 26,630,798,955
|
IssuesEvent
|
2023-01-24 17:40:40
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
30 ads history is blank
|
bug feature/rewards priority/P2 QA/Yes feature/ads OS/Android OS/Desktop
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
30 day ads history is blank.
Should probably determine if this link should be shown for "unverified" and "grandfather unverified" states in 1.48.x. If it is going to be shown, it should list the ads as it did previously.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Either be in unverified (new profile with 1.48.x) or grandfathered unverified (unverified upgraded from 1.47.x) state for Rewards or connected (to a custodian such as Uphold)
2. View at least one ad
3. Look at 30 day ads history on brave://rewards
## Actual result:
<!--Please add screenshots if needed-->
Ads history is blank
Unverified | Connected
---- | ----
<img width="1320" alt="Screen Shot 2023-01-23 at 4 00 44 PM" src="https://user-images.githubusercontent.com/28145373/214149131-38d9c2d6-4378-4dae-9210-b379e182a9ef.png"> | <img width="1312" alt="Screen Shot 2023-01-23 at 4 06 36 PM" src="https://user-images.githubusercontent.com/28145373/214150094-faea0ed9-ec9b-454d-b5c1-8efa8ffacd76.png">
## Expected result:
TBD - maybe not show this link if a user isn't earning? @Miyayes and someone from ads team should determine this.
If the link is shown though, ads information should be listed on the history.
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
1.48.x
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? 1.47.x no
- Can you reproduce this issue with the beta channel? 1.48.x yes
- Can you reproduce this issue with the nightly channel? 1.49.x yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc @Miyayes @tmancey @btlechowski
|
1.0
|
30 ads history is blank - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
30 day ads history is blank.
Should probably determine if this link should be shown for "unverified" and "grandfather unverified" states in 1.48.x. If it is going to be shown, it should list the ads as it did previously.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Either be in unverified (new profile with 1.48.x) or grandfathered unverified (unverified upgraded from 1.47.x) state for Rewards or connected (to a custodian such as Uphold)
2. View at least one ad
3. Look at 30 day ads history on brave://rewards
## Actual result:
<!--Please add screenshots if needed-->
Ads history is blank
Unverified | Connected
---- | ----
<img width="1320" alt="Screen Shot 2023-01-23 at 4 00 44 PM" src="https://user-images.githubusercontent.com/28145373/214149131-38d9c2d6-4378-4dae-9210-b379e182a9ef.png"> | <img width="1312" alt="Screen Shot 2023-01-23 at 4 06 36 PM" src="https://user-images.githubusercontent.com/28145373/214150094-faea0ed9-ec9b-454d-b5c1-8efa8ffacd76.png">
## Expected result:
TBD - maybe not show this link if a user isn't earning? @Miyayes and someone from ads team should determine this.
If the link is shown though, ads information should be listed on the history.
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
1.48.x
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? 1.47.x no
- Can you reproduce this issue with the beta channel? 1.48.x yes
- Can you reproduce this issue with the nightly channel? 1.49.x yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc @Miyayes @tmancey @btlechowski
|
non_process
|
ads history is blank have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description day ads history is blank should probably determine if this link should be shown for unverified and grandfather unverified states in x if it is going to be shown it should list the ads as it did previously steps to reproduce either be in unverified new profile with x or grandfathered unverified unverified upgraded from x state for rewards or connected to a custodian such as uphold view at least one ad look at day ads history on brave rewards actual result ads history is blank unverified connected img width alt screen shot at pm src img width alt screen shot at pm src expected result tbd maybe not show this link if a user isn t earning miyayes and someone from ads team should determine this if the link is shown though ads information should be listed on the history reproduces how often easily brave version brave version info x version channel information can you reproduce this issue with the current release x no can you reproduce this issue with the beta channel x yes can you reproduce this issue with the nightly channel x yes other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information cc miyayes tmancey btlechowski
| 0
|
324,830
| 9,913,197,010
|
IssuesEvent
|
2019-06-28 11:03:22
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
mobile.twitter.com - see bug description
|
browser-fenix engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://mobile.twitter.com/home
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: LastPass doesn't recognise site
**Steps to Reproduce**:
Tried to use LastPass to login, it thought I was logging in to Firefox website.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
mobile.twitter.com - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://mobile.twitter.com/home
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: LastPass doesn't recognise site
**Steps to Reproduce**:
Tried to use LastPass to login, it thought I was logging in to Firefox website.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
mobile twitter com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description lastpass doesn t recognise site steps to reproduce tried to use lastpass to login it thought i was logging in to firefox website browser configuration none from with ❤️
| 0
|
1,527
| 4,118,762,870
|
IssuesEvent
|
2016-06-08 12:48:44
|
World4Fly/Interface-for-Arduino
|
https://api.github.com/repos/World4Fly/Interface-for-Arduino
|
closed
|
Design a tester
|
process
|
Create basic concept of the hardware tester _considering_ following points:
- demonstrability
- usability for different purposes (game controller, input device)
- parts required
- costs
- visual design
- time required
|
1.0
|
Design a tester - Create basic concept of the hardware tester _considering_ following points:
- demonstrability
- usability for different purposes (game controller, input device)
- parts required
- costs
- visual design
- time required
|
process
|
design a tester create basic concept of the hardware tester considering following points demonstrability usability for different purposes game controller input device parts required costs visual design time required
| 1
|
559
| 2,502,400,271
|
IssuesEvent
|
2015-01-09 08:29:56
|
fossology/fossology
|
https://api.github.com/repos/fossology/fossology
|
opened
|
License browser regression - schedule link
|
Category: UI Component: Rank Component: Tester Priority: High Status: New Tracker: Bug
|
---
Author Name: **Bob Gobeille**
Original Redmine Issue: 7539, http://www.fossology.org/issues/7539
Original Date: 2014/08/26
Original Assignee: larry shi
---
If you pull up the license browser and there is no license scan data, you used to get a message "No data available" and a link to schedule a scan. Look at the bucket and copyright browsers for examples. Now we only get a text message. The idea behind giving the user a link is to make this task much simpler for the user. The current text "There is no successful scan for this upload, please schedule one license scanner on this upload" doesn't even tell a user how to do this. And if they know how to do it, it takes several clicks.
|
1.0
|
License browser regression - schedule link - ---
Author Name: **Bob Gobeille**
Original Redmine Issue: 7539, http://www.fossology.org/issues/7539
Original Date: 2014/08/26
Original Assignee: larry shi
---
If you pull up the license browser and there is no license scan data, you used to get a message "No data available" and a link to schedule a scan. Look at the bucket and copyright browsers for examples. Now we only get a text message. The idea behind giving the user a link is to make this task much simpler for the user. The current text "There is no successful scan for this upload, please schedule one license scanner on this upload" doesn't even tell a user how to do this. And if they know how to do it, it takes several clicks.
|
non_process
|
license browser regression schedule link author name bob gobeille original redmine issue original date original assignee larry shi if you pull up the license browser and there is no license scan data you used to get a message no data available and a link to schedule a scan look at the bucket and copyright browsers for examples now we only get a text message the idea behind giving the user a link is to make this task much simpler for the user the current text there is no successful scan for this upload please schedule one license scanner on this upload doesn t even tell a user how to do this and if they know how to do it it takes several clicks
| 0
|
275,365
| 8,575,605,455
|
IssuesEvent
|
2018-11-12 17:45:20
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
Enzo reader failing on windows
|
Bug Likelihood: 3 - Occasional Priority: Normal Severity: 4 - Crash / Wrong Results
|
Jean Favre reported that his Enzo files could not be read on Windows.
There are full-path specifications in his .hierarchy and .boundary files.
The issue may be related to parsing those paths expecting windows-style path separators (since running on Windows), but the files contain unix-style path separators.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 799
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Enzo reader failing on windows
Assigned to: Kathleen Biagas
Category:
Target version: 2.3.2
Author: Kathleen Biagas
Start: 07/27/2011
Due date:
% Done: 0
Estimated time: 2.0
Created: 07/27/2011 01:43 pm
Updated: 07/28/2011 01:42 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.3.0
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
Jean Favre reported that his Enzo files could not be read on Windows.
There are full-path specifications in his .hierarchy and .boundary files.
The issue may be related to parsing those paths expecting windows-style path separators (since running on Windows), but the files contain unix-style path separators.
Comments:
The Enzo reader uses StringHelpers::Basename to parse paths contained in the file. The method searched for sys-dependent path delimiters, and failed when Enzo files generated on *nix were read on Windows. Changed the StringHelpers methods to look for both unix-style and windows-style path delimiters.M /src/common/utility/StringHelpers.C2.3 SVN Revision 15587trunk SVN Revision 15589
|
1.0
|
Enzo reader failing on windows - Jean Favre reported that his Enzo files could not be read on Windows.
There are full-path specifications in his .hierarchy and .boundary files.
The issue may be related to parsing those paths expecting windows-style path separators (since running on Windows), but the files contain unix-style path separators.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 799
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Enzo reader failing on windows
Assigned to: Kathleen Biagas
Category:
Target version: 2.3.2
Author: Kathleen Biagas
Start: 07/27/2011
Due date:
% Done: 0
Estimated time: 2.0
Created: 07/27/2011 01:43 pm
Updated: 07/28/2011 01:42 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.3.0
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
Jean Favre reported that his Enzo files could not be read on Windows.
There are full-path specifications in his .hierarchy and .boundary files.
The issue may be related to parsing those paths expecting windows-style path separators (since running on Windows), but the files contain unix-style path separators.
Comments:
The Enzo reader uses StringHelpers::Basename to parse paths contained in the file. The method searched for sys-dependent path delimiters, and failed when Enzo files generated on *nix were read on Windows. Changed the StringHelpers methods to look for both unix-style and windows-style path delimiters.M /src/common/utility/StringHelpers.C2.3 SVN Revision 15587trunk SVN Revision 15589
|
non_process
|
enzo reader failing on windows jean favre reported that his enzo files could not be read on windows there are full path specifications in his hierarchy and boundary files the issue may be related to parsing those paths expecting windows style path separators since running on windows but the files contain unix style path separators redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject enzo reader failing on windows assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os windows support group any description jean favre reported that his enzo files could not be read on windows there are full path specifications in his hierarchy and boundary files the issue may be related to parsing those paths expecting windows style path separators since running on windows but the files contain unix style path separators comments the enzo reader uses stringhelpers basename to parse paths contained in the file the method searched for sys dependent path delimiters and failed when enzo files generated on nix were read on windows changed the stringhelpers methods to look for both unix style and windows style path delimiters m src common utility stringhelpers svn revision svn revision
| 0
|
360
| 2,795,024,911
|
IssuesEvent
|
2015-05-11 19:47:50
|
scieloorg/search-journals
|
https://api.github.com/repos/scieloorg/search-journals
|
closed
|
Cluster Coleção deve ser classificado pelo nome do país
|
Processamento
|
- Classificar pelo nome do pais
- O ideal seria ter dois botões para indicar a classificação desejada
|
1.0
|
Cluster Coleção deve ser classificado pelo nome do país - - Classificar pelo nome do pais
- O ideal seria ter dois botões para indicar a classificação desejada
|
process
|
cluster coleção deve ser classificado pelo nome do país classificar pelo nome do pais o ideal seria ter dois botões para indicar a classificação desejada
| 1
|
55,947
| 8,036,454,923
|
IssuesEvent
|
2018-07-30 09:12:20
|
fossasia/susi_linux
|
https://api.github.com/repos/fossasia/susi_linux
|
closed
|
To Add documentation and installation regarding the ReSpeaker
|
documentation
|
Current Behavior
No documentation regarding the ReSpeaker present
Expected Behaviour
To add the respected documentation
Would you like to work on this issue?
Yes
|
1.0
|
To Add documentation and installation regarding the ReSpeaker - Current Behavior
No documentation regarding the ReSpeaker present
Expected Behaviour
To add the respected documentation
Would you like to work on this issue?
Yes
|
non_process
|
to add documentation and installation regarding the respeaker current behavior no documentation regarding the respeaker present expected behaviour to add the respected documentation would you like to work on this issue yes
| 0
|
12,484
| 14,951,331,875
|
IssuesEvent
|
2021-01-26 14:16:28
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
add_docker_metadata false positive
|
:Processors Journalbeat Stalled Team:Integrations [zube]: Old-Inbox bug containers
|
The following log entry is populated erroneously with docker metadata.
My assumption is that journalbeat looks at the unit and finds "docker" in the string and therefore assumes it is log entry from inside a container. There is really no need to do that logic if journald is used as a log driver in docker it will say explicitly what logs that comes from a container. Just expose the information that is there. Example attached in the end.
Journal entry:
```
{
"_EXE" : "/usr/lib/systemd/systemd",
"MESSAGE" : "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount: Succeeded.",
"_PID" : "1",
"_GID" : "0",
"__CURSOR" : "s=9d5f42767d6641ed88fd402aed15dfca;i=905015;b=76232f86194c4353be9a3b46aee4e60b;m=255e5f3d4;t=584ee6c5c1f77;x=9345d20b3d6610f4",
"_HOSTNAME" : "puff",
"_UID" : "0",
"SYSLOG_FACILITY" : "3",
"__REALTIME_TIMESTAMP" : "1553534438612855",
"_SYSTEMD_SLICE" : "-.slice",
"_MACHINE_ID" : "b52ee41f2ace4198adfe01a971521a0d",
"PRIORITY" : "6",
"_SYSTEMD_CGROUP" : "/init.scope",
"MESSAGE_ID" : "7ad2d189f7e94e70a38c781354912448",
"_CMDLINE" : "/sbin/init",
"_COMM" : "systemd",
"_SOURCE_REALTIME_TIMESTAMP" : "1553534438612833",
"_SYSTEMD_UNIT" : "init.scope",
"_BOOT_ID" : "76232f86194c4353be9a3b46aee4e60b",
"__MONOTONIC_TIMESTAMP" : "10031068116",
"UNIT" : "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount",
"INVOCATION_ID" : "d3d87e303b21484db60bba05e2631595",
"CODE_FUNC" : "unit_log_success",
"CODE_FILE" : "../systemd-stable/src/core/unit.c",
"_TRANSPORT" : "journal",
"CODE_LINE" : "5496",
"SYSLOG_IDENTIFIER" : "systemd",
"_CAP_EFFECTIVE" : "3fffffffff"
}
```
The complete output in ES is as follows
```
{
"_index": "journalbeat-6.7.1-2019.03.25",
"_type": "doc",
"_id": "jGih9GkBvLolRgt0LWNc",
"_version": 1,
"_score": null,
"_source": {
"@timestamp": "2019-03-25T17:20:38.612Z",
"systemd": {
"unit": "init.scope",
"cgroup": "/init.scope",
"slice": "-.slice",
"transport": "journal"
},
"beat": {
"version": "6.7.1",
"name": "puff",
"hostname": "puff"
},
"journald": {
"code": {
"file": "../systemd-stable/src/core/unit.c",
"line": 5496,
"func": "unit_log_success"
}
},
"read_timestamp": "2019-04-06T21:49:55.983Z",
"process": {
"pid": 1,
"executable": "/usr/lib/systemd/systemd",
"name": "systemd",
"uid": 0,
"capabilites": "3fffffffff",
"cmd": "/sbin/init"
},
"syslog": {
"identifier": "systemd",
"priority": 6,
"facility": 3
},
"docker": {
"container": {
"id": "a7e942190a47f3997ce8e6e70afaf454e035f6692e7bdd536bfab6f11262224b",
"labels": {
"org": {
"label-schema": {
"build-date": "20190305",
"schema-version": "1.0",
"name": "journalbeat",
"license": "GPLv2",
"vcs-ref": "1d55b4bd9dbf106a4ad4bc34fe9ee425d922363b",
"vcs-url": "github.com/elastic/beats",
"version": "6.7.1",
"vendor": "Elastic",
"url": "https://www.elastic.co/products/beats/journalbeat"
}
},
"com": {
"docker": {
"compose": {
"project": "log",
"version": "1.24.0",
"oneoff": "False",
"config-hash": "5b98565d425a986b55aa1b95164247391612c2ab8c2b0bcd3b69cc882d1cce83",
"container-number": "1",
"service": "journalbeat"
}
}
},
"description": "Journalbeat ships systemd journal entries to Elasticsearch or Logstash.",
"license": "Elastic License"
},
"image": "docker.elastic.co/beats/journalbeat:6.7.1",
"name": "log_journalbeat_1"
}
},
"host": {
"id": "b52ee41f2ace4198adfe01a971521a0d",
"ip": [
"192.168.86.111",
"fe80::2aaa:e66e:cab0:8353",
"192.168.1.5",
"172.18.0.1",
"169.254.183.107",
"fe80::42:2aff:fe23:9fb6",
"172.19.0.1",
"169.254.144.12",
"fe80::1cbe:dfff:fe11:685b",
"169.254.160.155",
"fe80::4c49:51ff:fe59:faa2"
],
"mac": [
"38:d5:47:aa:c1:2a",
"02:42:d5:1a:1e:e5",
"02:42:2a:23:9f:b6",
"02:42:7b:1c:68:ce",
"1e:be:df:11:68:5b",
"4e:49:51:59:fa:a2"
],
"architecture": "x86_64",
"os": {
"version": "",
"family": "",
"name": "Arch Linux",
"build": "rolling",
"platform": "arch"
},
"containerized": true,
"boot_id": "76232f86194c4353be9a3b46aee4e60b",
"name": "puff"
},
"message": "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount: Succeeded.",
"custom": {
"unit": "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount",
"message_id": "7ad2d189f7e94e70a38c781354912448",
"invocation_id": "d3d87e303b21484db60bba05e2631595"
}
},
"fields": {
"@timestamp": [
"2019-03-25T17:20:38.612Z"
]
},
"highlight": {
"message": [
"@kibana-highlighted-field@var@/kibana-highlighted-field@-@kibana-highlighted-field@lib@/kibana-highlighted-field@-@kibana-highlighted-field@docker@/kibana-highlighted-field@-@kibana-highlighted-field@containers@/kibana-highlighted-field@-@kibana-highlighted-field@a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe@/kibana-highlighted-field@-@kibana-highlighted-field@mounts@/kibana-highlighted-field@-@kibana-highlighted-field@shm.mount@/kibana-highlighted-field@: @kibana-highlighted-field@Succeeded@/kibana-highlighted-field@."
]
},
"sort": [
1553534438612
]
}
```
Example journal entry from a container:
```
{
"_EXE" : "/usr/bin/dockerd",
"CONTAINER_TAG" : "a7e942190a47",
"_TRANSPORT" : "journal",
"_MACHINE_ID" : "b52ee41f2ace4198adfe01a971521a0d",
"_PID" : "7893",
"_SYSTEMD_CGROUP" : "/system.slice/docker.service",
"__CURSOR" : "s=9d5f42767d6641ed88fd402aed15dfca;i=c70ae1;b=a820b511fd2a434b87a5e27dda81ffb4;m=82fd8b26c;t=585e3b983439a;x=267490a0d8cf7a41",
"_SYSTEMD_INVOCATION_ID" : "6e8c014e4f104ea5942dca15add5d9fb",
"SYSLOG_IDENTIFIER" : "a7e942190a47",
"_CAP_EFFECTIVE" : "3fffffffff",
"MESSAGE" : "2019-04-06T22:00:00.010Z\tINFO\t[input]\tinput/input.go:133\tjournalbeat successfully published 50 events\t{\"id\": \"e5cc492d-ddd8->
"CONTAINER_ID_FULL" : "a7e942190a47f3997ce8e6e70afaf454e035f6692e7bdd536bfab6f11262224b",
"_BOOT_ID" : "a820b511fd2a434b87a5e27dda81ffb4",
"_HOSTNAME" : "puff",
"__REALTIME_TIMESTAMP" : "1554588000011162",
"CONTAINER_NAME" : "log_journalbeat_1",
"CONTAINER_ID" : "a7e942190a47",
"__MONOTONIC_TIMESTAMP" : "35162468972",
"_COMM" : "dockerd",
"_SYSTEMD_UNIT" : "docker.service",
"PRIORITY" : "3",
"_UID" : "0",
"_GID" : "0",
"_SOURCE_REALTIME_TIMESTAMP" : "1554588000011115",
"_CMDLINE" : "/usr/bin/dockerd -H fd:// --default-runtime=nvidia --node-generic-resource gpu=GPU-9cbe0570",
"_SYSTEMD_SLICE" : "system.slice"
}
```
|
1.0
|
add_docker_metadata false positive - The following log entry is populated erroneously with docker metadata.
My assumption is that journalbeat looks at the unit and finds "docker" in the string and therefore assumes it is log entry from inside a container. There is really no need to do that logic if journald is used as a log driver in docker it will say explicitly what logs that comes from a container. Just expose the information that is there. Example attached in the end.
Journal entry:
```
{
"_EXE" : "/usr/lib/systemd/systemd",
"MESSAGE" : "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount: Succeeded.",
"_PID" : "1",
"_GID" : "0",
"__CURSOR" : "s=9d5f42767d6641ed88fd402aed15dfca;i=905015;b=76232f86194c4353be9a3b46aee4e60b;m=255e5f3d4;t=584ee6c5c1f77;x=9345d20b3d6610f4",
"_HOSTNAME" : "puff",
"_UID" : "0",
"SYSLOG_FACILITY" : "3",
"__REALTIME_TIMESTAMP" : "1553534438612855",
"_SYSTEMD_SLICE" : "-.slice",
"_MACHINE_ID" : "b52ee41f2ace4198adfe01a971521a0d",
"PRIORITY" : "6",
"_SYSTEMD_CGROUP" : "/init.scope",
"MESSAGE_ID" : "7ad2d189f7e94e70a38c781354912448",
"_CMDLINE" : "/sbin/init",
"_COMM" : "systemd",
"_SOURCE_REALTIME_TIMESTAMP" : "1553534438612833",
"_SYSTEMD_UNIT" : "init.scope",
"_BOOT_ID" : "76232f86194c4353be9a3b46aee4e60b",
"__MONOTONIC_TIMESTAMP" : "10031068116",
"UNIT" : "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount",
"INVOCATION_ID" : "d3d87e303b21484db60bba05e2631595",
"CODE_FUNC" : "unit_log_success",
"CODE_FILE" : "../systemd-stable/src/core/unit.c",
"_TRANSPORT" : "journal",
"CODE_LINE" : "5496",
"SYSLOG_IDENTIFIER" : "systemd",
"_CAP_EFFECTIVE" : "3fffffffff"
}
```
The complete output in ES is as follows
```
{
"_index": "journalbeat-6.7.1-2019.03.25",
"_type": "doc",
"_id": "jGih9GkBvLolRgt0LWNc",
"_version": 1,
"_score": null,
"_source": {
"@timestamp": "2019-03-25T17:20:38.612Z",
"systemd": {
"unit": "init.scope",
"cgroup": "/init.scope",
"slice": "-.slice",
"transport": "journal"
},
"beat": {
"version": "6.7.1",
"name": "puff",
"hostname": "puff"
},
"journald": {
"code": {
"file": "../systemd-stable/src/core/unit.c",
"line": 5496,
"func": "unit_log_success"
}
},
"read_timestamp": "2019-04-06T21:49:55.983Z",
"process": {
"pid": 1,
"executable": "/usr/lib/systemd/systemd",
"name": "systemd",
"uid": 0,
"capabilites": "3fffffffff",
"cmd": "/sbin/init"
},
"syslog": {
"identifier": "systemd",
"priority": 6,
"facility": 3
},
"docker": {
"container": {
"id": "a7e942190a47f3997ce8e6e70afaf454e035f6692e7bdd536bfab6f11262224b",
"labels": {
"org": {
"label-schema": {
"build-date": "20190305",
"schema-version": "1.0",
"name": "journalbeat",
"license": "GPLv2",
"vcs-ref": "1d55b4bd9dbf106a4ad4bc34fe9ee425d922363b",
"vcs-url": "github.com/elastic/beats",
"version": "6.7.1",
"vendor": "Elastic",
"url": "https://www.elastic.co/products/beats/journalbeat"
}
},
"com": {
"docker": {
"compose": {
"project": "log",
"version": "1.24.0",
"oneoff": "False",
"config-hash": "5b98565d425a986b55aa1b95164247391612c2ab8c2b0bcd3b69cc882d1cce83",
"container-number": "1",
"service": "journalbeat"
}
}
},
"description": "Journalbeat ships systemd journal entries to Elasticsearch or Logstash.",
"license": "Elastic License"
},
"image": "docker.elastic.co/beats/journalbeat:6.7.1",
"name": "log_journalbeat_1"
}
},
"host": {
"id": "b52ee41f2ace4198adfe01a971521a0d",
"ip": [
"192.168.86.111",
"fe80::2aaa:e66e:cab0:8353",
"192.168.1.5",
"172.18.0.1",
"169.254.183.107",
"fe80::42:2aff:fe23:9fb6",
"172.19.0.1",
"169.254.144.12",
"fe80::1cbe:dfff:fe11:685b",
"169.254.160.155",
"fe80::4c49:51ff:fe59:faa2"
],
"mac": [
"38:d5:47:aa:c1:2a",
"02:42:d5:1a:1e:e5",
"02:42:2a:23:9f:b6",
"02:42:7b:1c:68:ce",
"1e:be:df:11:68:5b",
"4e:49:51:59:fa:a2"
],
"architecture": "x86_64",
"os": {
"version": "",
"family": "",
"name": "Arch Linux",
"build": "rolling",
"platform": "arch"
},
"containerized": true,
"boot_id": "76232f86194c4353be9a3b46aee4e60b",
"name": "puff"
},
"message": "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount: Succeeded.",
"custom": {
"unit": "var-lib-docker-containers-a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe-mounts-shm.mount",
"message_id": "7ad2d189f7e94e70a38c781354912448",
"invocation_id": "d3d87e303b21484db60bba05e2631595"
}
},
"fields": {
"@timestamp": [
"2019-03-25T17:20:38.612Z"
]
},
"highlight": {
"message": [
"@kibana-highlighted-field@var@/kibana-highlighted-field@-@kibana-highlighted-field@lib@/kibana-highlighted-field@-@kibana-highlighted-field@docker@/kibana-highlighted-field@-@kibana-highlighted-field@containers@/kibana-highlighted-field@-@kibana-highlighted-field@a34010a7c0ac153e46318ba05e865a7753a6a6523f939011f132b29cc7d021fe@/kibana-highlighted-field@-@kibana-highlighted-field@mounts@/kibana-highlighted-field@-@kibana-highlighted-field@shm.mount@/kibana-highlighted-field@: @kibana-highlighted-field@Succeeded@/kibana-highlighted-field@."
]
},
"sort": [
1553534438612
]
}
```
Example journal entry from a container:
```
{
"_EXE" : "/usr/bin/dockerd",
"CONTAINER_TAG" : "a7e942190a47",
"_TRANSPORT" : "journal",
"_MACHINE_ID" : "b52ee41f2ace4198adfe01a971521a0d",
"_PID" : "7893",
"_SYSTEMD_CGROUP" : "/system.slice/docker.service",
"__CURSOR" : "s=9d5f42767d6641ed88fd402aed15dfca;i=c70ae1;b=a820b511fd2a434b87a5e27dda81ffb4;m=82fd8b26c;t=585e3b983439a;x=267490a0d8cf7a41",
"_SYSTEMD_INVOCATION_ID" : "6e8c014e4f104ea5942dca15add5d9fb",
"SYSLOG_IDENTIFIER" : "a7e942190a47",
"_CAP_EFFECTIVE" : "3fffffffff",
"MESSAGE" : "2019-04-06T22:00:00.010Z\tINFO\t[input]\tinput/input.go:133\tjournalbeat successfully published 50 events\t{\"id\": \"e5cc492d-ddd8->
"CONTAINER_ID_FULL" : "a7e942190a47f3997ce8e6e70afaf454e035f6692e7bdd536bfab6f11262224b",
"_BOOT_ID" : "a820b511fd2a434b87a5e27dda81ffb4",
"_HOSTNAME" : "puff",
"__REALTIME_TIMESTAMP" : "1554588000011162",
"CONTAINER_NAME" : "log_journalbeat_1",
"CONTAINER_ID" : "a7e942190a47",
"__MONOTONIC_TIMESTAMP" : "35162468972",
"_COMM" : "dockerd",
"_SYSTEMD_UNIT" : "docker.service",
"PRIORITY" : "3",
"_UID" : "0",
"_GID" : "0",
"_SOURCE_REALTIME_TIMESTAMP" : "1554588000011115",
"_CMDLINE" : "/usr/bin/dockerd -H fd:// --default-runtime=nvidia --node-generic-resource gpu=GPU-9cbe0570",
"_SYSTEMD_SLICE" : "system.slice"
}
```
|
process
|
add docker metadata false positive the following log entry is populated erroneously with docker metadata my assumption is that journalbeat looks at the unit and finds docker in the string and therefore assumes it is log entry from inside a container there is really no need to do that logic if journald is used as a log driver in docker it will say explicitly what logs that comes from a container just expose the information that is there example attached in the end journal entry exe usr lib systemd systemd message var lib docker containers mounts shm mount succeeded pid gid cursor s i b m t x hostname puff uid syslog facility realtime timestamp systemd slice slice machine id priority systemd cgroup init scope message id cmdline sbin init comm systemd source realtime timestamp systemd unit init scope boot id monotonic timestamp unit var lib docker containers mounts shm mount invocation id code func unit log success code file systemd stable src core unit c transport journal code line syslog identifier systemd cap effective the complete output in es is as follows index journalbeat type doc id version score null source timestamp systemd unit init scope cgroup init scope slice slice transport journal beat version name puff hostname puff journald code file systemd stable src core unit c line func unit log success read timestamp process pid executable usr lib systemd systemd name systemd uid capabilites cmd sbin init syslog identifier systemd priority facility docker container id labels org label schema build date schema version name journalbeat license vcs ref vcs url github com elastic beats version vendor elastic url com docker compose project log version oneoff false config hash container number service journalbeat description journalbeat ships systemd journal entries to elasticsearch or logstash license elastic license image docker elastic co beats journalbeat name log journalbeat host id ip dfff mac aa ce be df fa architecture os version family name arch linux build rolling platform arch containerized true boot id name puff message var lib docker containers mounts shm mount succeeded custom unit var lib docker containers mounts shm mount message id invocation id fields timestamp highlight message kibana highlighted field var kibana highlighted field kibana highlighted field lib kibana highlighted field kibana highlighted field docker kibana highlighted field kibana highlighted field containers kibana highlighted field kibana highlighted field kibana highlighted field kibana highlighted field mounts kibana highlighted field kibana highlighted field shm mount kibana highlighted field kibana highlighted field succeeded kibana highlighted field sort example journal entry from a container exe usr bin dockerd container tag transport journal machine id pid systemd cgroup system slice docker service cursor s i b m t x systemd invocation id syslog identifier cap effective message tinfo t tinput input go tjournalbeat successfully published events t id container id full boot id hostname puff realtime timestamp container name log journalbeat container id monotonic timestamp comm dockerd systemd unit docker service priority uid gid source realtime timestamp cmdline usr bin dockerd h fd default runtime nvidia node generic resource gpu gpu systemd slice system slice
| 1
|
13,235
| 15,706,140,470
|
IssuesEvent
|
2021-03-26 17:01:58
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
display-buffer-alist reusable-frames does not work as expected
|
process:windows
|
This was originally posted on [emacs so](https://emacs.stackexchange.com/q/55556/21814).
When using the `display-buffer-alist` value from [ess doc](https://github.com/emacs-ess/ESS/blob/master/doc/ess.texi#L632-L642)
```lisp
(setq display-buffer-alist
'(("*R"
(display-buffer-reuse-window display-buffer-pop-up-frame)
(reusable-frames . 0))))
```
and starting R shell from some `ess-r-mode` buffer the following occurs: after creating a new frame displaying the R shell, Emacs displays a second instance of the `ess-r-mode` buffer in this new frame instead of returning to the old instance.
Everything works correctly when I set `reusable-frames` globally
```lisp
(setq display-buffer-reuse-frames 0)
```
but this should not be necessary.
Quoting [so answer](https://emacs.stackexchange.com/a/55571/21814):
The erroneous behaviour happens only the first time one types `C-c C-z` (`ess-switch-to-inferior-or-script-buffer`) in an `ess-r-mode` buffer, i.e. before the corresponding inferior `*R*` process buffer has been created.
This is because `ess-force-buffer-current` calls `ess-request-a-process` with a non-`nil` `noswitch` argument when an inferior process does not already exist.
`ess-request-a-process`, in turn, does the following as its last step:
```lisp
(if noswitch
(pop-to-buffer (current-buffer)) ;; VS: this is weird, but is necessary
(pop-to-buffer (buffer-name (process-buffer (get-process proc)))))
```
This call to `(pop-to-buffer (current-buffer))` is wrong and is what causes your `ess-r-mode` buffer to appear a second time in the new frame. It was added in the following commit from 2012: <https://github.com/emacs-ess/ESS/commit/b29ea8f934f7c08a512c73f14e914bca7229b3c1>
I boldly say it is wrong because popping to the current buffer is quite an intrusive operation (as indicated by the bug in question), and the original intention of the author can almost definitely be written in a better way. I don't know what issue the author originally faced, but perhaps the ESS devs can figure it out.
|
1.0
|
display-buffer-alist reusable-frames does not work as expected - This was originally posted on [emacs so](https://emacs.stackexchange.com/q/55556/21814).
When using the `display-buffer-alist` value from [ess doc](https://github.com/emacs-ess/ESS/blob/master/doc/ess.texi#L632-L642)
```lisp
(setq display-buffer-alist
'(("*R"
(display-buffer-reuse-window display-buffer-pop-up-frame)
(reusable-frames . 0))))
```
and starting R shell from some `ess-r-mode` buffer the following occurs: after creating a new frame displaying the R shell, Emacs displays a second instance of the `ess-r-mode` buffer in this new frame instead of returning to the old instance.
Everything works correctly when I set `reusable-frames` globally
```lisp
(setq display-buffer-reuse-frames 0)
```
but this should not be necessary.
Quoting [so answer](https://emacs.stackexchange.com/a/55571/21814):
The erroneous behaviour happens only the first time one types `C-c C-z` (`ess-switch-to-inferior-or-script-buffer`) in an `ess-r-mode` buffer, i.e. before the corresponding inferior `*R*` process buffer has been created.
This is because `ess-force-buffer-current` calls `ess-request-a-process` with a non-`nil` `noswitch` argument when an inferior process does not already exist.
`ess-request-a-process`, in turn, does the following as its last step:
```lisp
(if noswitch
(pop-to-buffer (current-buffer)) ;; VS: this is weird, but is necessary
(pop-to-buffer (buffer-name (process-buffer (get-process proc)))))
```
This call to `(pop-to-buffer (current-buffer))` is wrong and is what causes your `ess-r-mode` buffer to appear a second time in the new frame. It was added in the following commit from 2012: <https://github.com/emacs-ess/ESS/commit/b29ea8f934f7c08a512c73f14e914bca7229b3c1>
I boldly say it is wrong because popping to the current buffer is quite an intrusive operation (as indicated by the bug in question), and the original intention of the author can almost definitely be written in a better way. I don't know what issue the author originally faced, but perhaps the ESS devs can figure it out.
|
process
|
display buffer alist reusable frames does not work as expected this was originally posted on when using the display buffer alist value from lisp setq display buffer alist r display buffer reuse window display buffer pop up frame reusable frames and starting r shell from some ess r mode buffer the following occurs after creating a new frame displaying the r shell emacs displays a second instance of the ess r mode buffer in this new frame instead of returning to the old instance everything works correctly when i set reusable frames globally lisp setq display buffer reuse frames but this should not be necessary quoting the erroneous behaviour happens only the first time one types c c c z ess switch to inferior or script buffer in an ess r mode buffer i e before the corresponding inferior r process buffer has been created this is because ess force buffer current calls ess request a process with a non nil noswitch argument when an inferior process does not already exist ess request a process in turn does the following as its last step lisp if noswitch pop to buffer current buffer vs this is weird but is necessary pop to buffer buffer name process buffer get process proc this call to pop to buffer current buffer is wrong and is what causes your ess r mode buffer to appear a second time in the new frame it was added in the following commit from i boldly say it is wrong because popping to the current buffer is quite an intrusive operation as indicated by the bug in question and the original intention of the author can almost definitely be written in a better way i don t know what issue the author originally faced but perhaps the ess devs can figure it out
| 1
|
15,424
| 19,609,402,595
|
IssuesEvent
|
2022-01-06 13:43:04
|
Kernem/FeRSS-Core
|
https://api.github.com/repos/Kernem/FeRSS-Core
|
opened
|
ChannelCollection might need to own channels
|
fetching post-processing
|
If ChannelCollections are to persist across draw functions it needs to own it's channels. Te alternative is to create a new collection each time the UI needs to be drawn, which would necessitate fetching all the content over and over again.
|
1.0
|
ChannelCollection might need to own channels - If ChannelCollections are to persist across draw functions it needs to own it's channels. Te alternative is to create a new collection each time the UI needs to be drawn, which would necessitate fetching all the content over and over again.
|
process
|
channelcollection might need to own channels if channelcollections are to persist across draw functions it needs to own it s channels te alternative is to create a new collection each time the ui needs to be drawn which would necessitate fetching all the content over and over again
| 1
|
1,537
| 4,151,045,464
|
IssuesEvent
|
2016-06-15 19:17:50
|
Truncatis-Life/Server_Files
|
https://api.github.com/repos/Truncatis-Life/Server_Files
|
opened
|
Standard Server files changing
|
enhancement Processing
|
description.ext
briefing.sql
serverpic
etc
etc
etc
etc
etc
you get the point
|
1.0
|
Standard Server files changing - description.ext
briefing.sql
serverpic
etc
etc
etc
etc
etc
you get the point
|
process
|
standard server files changing description ext briefing sql serverpic etc etc etc etc etc you get the point
| 1
|
6,101
| 8,960,786,084
|
IssuesEvent
|
2019-01-28 07:31:22
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
opened
|
after creating a document, you can add the creator as a partner even though he's already added
|
2.0.7 Process bug
|
after creating a document, you can add the creator as a partner even though he's already added to the document when he was created

after adding the user again the changes are saved

|
1.0
|
after creating a document, you can add the creator as a partner even though he's already added - after creating a document, you can add the creator as a partner even though he's already added to the document when he was created

after adding the user again the changes are saved

|
process
|
after creating a document you can add the creator as a partner even though he s already added after creating a document you can add the creator as a partner even though he s already added to the document when he was created after adding the user again the changes are saved
| 1
|
12,846
| 15,228,060,240
|
IssuesEvent
|
2021-02-18 10:59:39
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Buffers from process.stdout are not transferable
|
buffer child_process worker
|
The buffers from process.stdout (child process) don't seem to be transferable. Is there a way to make them so?
@addaleax @jasnell
I'm trying to do something like:
```js
const proc = cp.spawn(...)
const parser = new ParserWorker()
proc.stdout.on('data', buf => parser.postMessage(buf, [ buf ]))
```
|
1.0
|
Buffers from process.stdout are not transferable - The buffers from process.stdout (child process) don't seem to be transferable. Is there a way to make them so?
@addaleax @jasnell
I'm trying to do something like:
```js
const proc = cp.spawn(...)
const parser = new ParserWorker()
proc.stdout.on('data', buf => parser.postMessage(buf, [ buf ]))
```
|
process
|
buffers from process stdout are not transferable the buffers from process stdout child process don t seem to be transferable is there a way to make them so addaleax jasnell i m trying to do something like js const proc cp spawn const parser new parserworker proc stdout on data buf parser postmessage buf
| 1
|
26,240
| 12,893,546,998
|
IssuesEvent
|
2020-07-13 21:53:11
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
https://github.com/flutter/flutter/pull/49708 had a dramatic impact on a benchmark
|
P6 engine framework perf: speed severe: performance severe: regression
|
The PR https://github.com/flutter/flutter/pull/49708 produced a change in the `flutter_gallery__transition_perf_with_semantics/delta_average_frame_build_time_millis` benchmark that tripled the time those frame build times take.
This impact was seen on 1/17/2020 when the original version of this fix landed under https://github.com/flutter/flutter/pull/48985
The benchmarks went back to their previous times when that PR was reverted on 1/27/2020 (see https://github.com/flutter/flutter/pull/49572)
The impact was again felt when the changes were re-landed on 1/29/2020 (see https://github.com/flutter/flutter/pull/49708)
|
True
|
https://github.com/flutter/flutter/pull/49708 had a dramatic impact on a benchmark - The PR https://github.com/flutter/flutter/pull/49708 produced a change in the `flutter_gallery__transition_perf_with_semantics/delta_average_frame_build_time_millis` benchmark that tripled the time those frame build times take.
This impact was seen on 1/17/2020 when the original version of this fix landed under https://github.com/flutter/flutter/pull/48985
The benchmarks went back to their previous times when that PR was reverted on 1/27/2020 (see https://github.com/flutter/flutter/pull/49572)
The impact was again felt when the changes were re-landed on 1/29/2020 (see https://github.com/flutter/flutter/pull/49708)
|
non_process
|
had a dramatic impact on a benchmark the pr produced a change in the flutter gallery transition perf with semantics delta average frame build time millis benchmark that tripled the time those frame build times take this impact was seen on when the original version of this fix landed under the benchmarks went back to their previous times when that pr was reverted on see the impact was again felt when the changes were re landed on see
| 0
|
13,114
| 5,300,798,607
|
IssuesEvent
|
2017-02-10 06:58:40
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
closed
|
Allow config to specify disk label for virtual floppy
|
builder/vmware-esxi enhancement waiting-reply
|
From reading the Redhat documentation, it appears that we can avoid doing the insecure VNC connection to launch a kickstart install if we use `floppy_files` AND label the FAT filesystem created `OEMDRV`. Right now, the disk label is hard-coded to "packer".
This enhancement request is to allow us to specify `"floppy_label"` in our configuration file to support VNC-free kickstart installs.
|
1.0
|
Allow config to specify disk label for virtual floppy - From reading the Redhat documentation, it appears that we can avoid doing the insecure VNC connection to launch a kickstart install if we use `floppy_files` AND label the FAT filesystem created `OEMDRV`. Right now, the disk label is hard-coded to "packer".
This enhancement request is to allow us to specify `"floppy_label"` in our configuration file to support VNC-free kickstart installs.
|
non_process
|
allow config to specify disk label for virtual floppy from reading the redhat documentation it appears that we can avoid doing the insecure vnc connection to launch a kickstart install if we use floppy files and label the fat filesystem created oemdrv right now the disk label is hard coded to packer this enhancement request is to allow us to specify floppy label in our configuration file to support vnc free kickstart installs
| 0
|
16,138
| 6,109,591,188
|
IssuesEvent
|
2017-06-21 13:25:41
|
mavlink/mavros
|
https://api.github.com/repos/mavlink/mavros
|
closed
|
mavros_extras error
|
build error question
|
I'm intalling ros on raspberry pi 3 model B and i get this error when building mavros_extras >>>
please help >>>>>>>>>>>>
pi@raspberrypi:~ $ cd /home/pi/ros_catkin_ws/build_isolated/mavros_extras && /opt/ros/indigo/env.sh make -j4 -l4
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge
//usr/usrbin//binld/:ld :cannot cannotopen open output file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/outputmavros_extras/copter_visualization: Is a directory
file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher: Is a directory
collect2: error: ld returned 1 exit status
collect2: error: ld returned 1 exit status
CMakeFiles/servo_state_publisher.dir/build.make:155: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher] Error 1
CMakeFiles/Makefile2:195: recipe for target 'CMakeFiles/servo_state_publisher.dir/all' failed
make[1]: *** [CMakeFiles/servo_state_publisher.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
CMakeFiles/copter_visualization.dir/build.make:155: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization] Error 1
CMakeFiles/Makefile2:717: recipe for target 'CMakeFiles/copter_visualization.dir/all' failed
make[1]: *** [CMakeFiles/copter_visualization.dir/all] Error 2
/usr/bin/ld: cannot open output file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge: Is a directory
collect2: error: ld returned 1 exit status
CMakeFiles/gcs_image_bridge.dir/build.make:228: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge] Error 1
CMakeFiles/Makefile2:552: recipe for target 'CMakeFiles/gcs_image_bridge.dir/all' failed
make[1]: *** [CMakeFiles/gcs_image_bridge.dir/all] Error 2
[ 80%] Built target mavros_extras
Makefile:138: recipe for target 'all' failed
make: *** [all] Error 2
|
1.0
|
mavros_extras error - I'm intalling ros on raspberry pi 3 model B and i get this error when building mavros_extras >>>
please help >>>>>>>>>>>>
pi@raspberrypi:~ $ cd /home/pi/ros_catkin_ws/build_isolated/mavros_extras && /opt/ros/indigo/env.sh make -j4 -l4
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization
[ 80%] Linking CXX executable /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge
//usr/usrbin//binld/:ld :cannot cannotopen open output file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/outputmavros_extras/copter_visualization: Is a directory
file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher: Is a directory
collect2: error: ld returned 1 exit status
collect2: error: ld returned 1 exit status
CMakeFiles/servo_state_publisher.dir/build.make:155: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/servo_state_publisher] Error 1
CMakeFiles/Makefile2:195: recipe for target 'CMakeFiles/servo_state_publisher.dir/all' failed
make[1]: *** [CMakeFiles/servo_state_publisher.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
CMakeFiles/copter_visualization.dir/build.make:155: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/copter_visualization] Error 1
CMakeFiles/Makefile2:717: recipe for target 'CMakeFiles/copter_visualization.dir/all' failed
make[1]: *** [CMakeFiles/copter_visualization.dir/all] Error 2
/usr/bin/ld: cannot open output file /home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge: Is a directory
collect2: error: ld returned 1 exit status
CMakeFiles/gcs_image_bridge.dir/build.make:228: recipe for target '/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge' failed
make[2]: *** [/home/pi/ros_catkin_ws/devel_isolated/mavros_extras/lib/mavros_extras/gcs_image_bridge] Error 1
CMakeFiles/Makefile2:552: recipe for target 'CMakeFiles/gcs_image_bridge.dir/all' failed
make[1]: *** [CMakeFiles/gcs_image_bridge.dir/all] Error 2
[ 80%] Built target mavros_extras
Makefile:138: recipe for target 'all' failed
make: *** [all] Error 2
|
non_process
|
mavros extras error i m intalling ros on raspberry pi model b and i get this error when building mavros extras please help pi raspberrypi cd home pi ros catkin ws build isolated mavros extras opt ros indigo env sh make linking cxx executable home pi ros catkin ws devel isolated mavros extras lib mavros extras servo state publisher linking cxx executable home pi ros catkin ws devel isolated mavros extras lib mavros extras copter visualization linking cxx executable home pi ros catkin ws devel isolated mavros extras lib mavros extras gcs image bridge usr usrbin binld ld cannot cannotopen open output file home pi ros catkin ws devel isolated mavros extras lib outputmavros extras copter visualization is a directory file home pi ros catkin ws devel isolated mavros extras lib mavros extras servo state publisher is a directory error ld returned exit status error ld returned exit status cmakefiles servo state publisher dir build make recipe for target home pi ros catkin ws devel isolated mavros extras lib mavros extras servo state publisher failed make error cmakefiles recipe for target cmakefiles servo state publisher dir all failed make error make waiting for unfinished jobs cmakefiles copter visualization dir build make recipe for target home pi ros catkin ws devel isolated mavros extras lib mavros extras copter visualization failed make error cmakefiles recipe for target cmakefiles copter visualization dir all failed make error usr bin ld cannot open output file home pi ros catkin ws devel isolated mavros extras lib mavros extras gcs image bridge is a directory error ld returned exit status cmakefiles gcs image bridge dir build make recipe for target home pi ros catkin ws devel isolated mavros extras lib mavros extras gcs image bridge failed make error cmakefiles recipe for target cmakefiles gcs image bridge dir all failed make error built target mavros extras makefile recipe for target all failed make error
| 0
|
43,203
| 17,466,875,854
|
IssuesEvent
|
2021-08-06 18:12:14
|
Azure/azure-sdk-for-net
|
https://api.github.com/repos/Azure/azure-sdk-for-net
|
closed
|
[Servicebus] test failing in nightly runs
|
Service Bus Client test-reliability
|
Servicebus nightly test runs are failing with:
> Error message
> Multiple failures or warnings in test:
>
> Azure.Messaging.ServiceBus.ServiceBusException: The session lock was lost. Request a new session receiver. (SessionLockLost) ---> Azure.Messaging.ServiceBus.ServiceBusException: The write operation failed, see inner exception. ErrorCode: ConnectionReset (ServiceCommunicationProblem) ---> System.IO.IOException: The write operation failed, see inner exception. ---> System.Net.Sockets.SocketException: An existing connection was forcibly closed by the remote host
> at Microsoft.Azure.Amqp.Transport.TransportStream.EndWrite(IAsyncResult asyncResult)
> at System.Net.Security.SslStream.StartWriting(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> at System.Net.Security.SslStream.ProcessWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> --- End of inner exception stack trace ---
> at System.Net.Security.SslStream.ProcessWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> at System.Net.Security.SslStream.BeginWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncCallback asyncCallback, Object asyncState)
> at Microsoft.Azure.Amqp.Transport.TlsTransport.WriteAsync(TransportAsyncCallbackArgs args)
> at Microsoft.Azure.Amqp.AsyncIO.AsyncWriter.WriteCore()
> --- End of inner exception stack trace ---
> --- End of inner exception stack trace ---
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.ThrowIfSessionLockLost() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1440
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<ExecuteRequest>d__59.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1102
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Runtime.CompilerServices.ConfiguredTaskAwaitable1.ConfiguredTaskAwaiter.GetResult() at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<GetStateInternal>d__63.MoveNext() in /_/sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1181 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.ConfiguredTaskAwaitable1.ConfiguredTaskAwaiter.GetResult()
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<>c.<<GetStateAsync>b__62_0>d.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1164
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Threading.Tasks.ValueTask1.get_Result() at Azure.Messaging.ServiceBus.ServiceBusRetryPolicy.<RunOperation>d__212.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Primitives/ServiceBusRetryPolicy.cs:line 144
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.ServiceBusRetryPolicy.<RunOperation>d__212.MoveNext() in /_/sdk/servicebus/Azure.Messaging.ServiceBus/src/Primitives/ServiceBusRetryPolicy.cs:line 164 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Threading.Tasks.ValueTask1.get_Result()
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<GetStateAsync>d__62.MoveNext
>
> Stack trace
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 176
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 187
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 187
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 188
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
> at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted()
> at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter)
> at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke)
> at NUnit.Framework.Internal.Commands.TestMethodCommand.RunTestMethod(TestExecutionContext context)
> at NUnit.Framework.Internal.Commands.TestMethodCommand.Execute(TestExecutionContext context)
> at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.<>c__DisplayClass1_0.<Execute>b__0()
> at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.RunTestMethodInThreadAbortSafeZone(TestExecutionContext context, Action action)
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1034081&view=results
|
1.0
|
[Servicebus] test failing in nightly runs - Servicebus nightly test runs are failing with:
> Error message
> Multiple failures or warnings in test:
>
> Azure.Messaging.ServiceBus.ServiceBusException: The session lock was lost. Request a new session receiver. (SessionLockLost) ---> Azure.Messaging.ServiceBus.ServiceBusException: The write operation failed, see inner exception. ErrorCode: ConnectionReset (ServiceCommunicationProblem) ---> System.IO.IOException: The write operation failed, see inner exception. ---> System.Net.Sockets.SocketException: An existing connection was forcibly closed by the remote host
> at Microsoft.Azure.Amqp.Transport.TransportStream.EndWrite(IAsyncResult asyncResult)
> at System.Net.Security.SslStream.StartWriting(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> at System.Net.Security.SslStream.ProcessWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> --- End of inner exception stack trace ---
> at System.Net.Security.SslStream.ProcessWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest)
> at System.Net.Security.SslStream.BeginWrite(Byte[] buffer, Int32 offset, Int32 count, AsyncCallback asyncCallback, Object asyncState)
> at Microsoft.Azure.Amqp.Transport.TlsTransport.WriteAsync(TransportAsyncCallbackArgs args)
> at Microsoft.Azure.Amqp.AsyncIO.AsyncWriter.WriteCore()
> --- End of inner exception stack trace ---
> --- End of inner exception stack trace ---
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.ThrowIfSessionLockLost() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1440
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<ExecuteRequest>d__59.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1102
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Runtime.CompilerServices.ConfiguredTaskAwaitable1.ConfiguredTaskAwaiter.GetResult() at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<GetStateInternal>d__63.MoveNext() in /_/sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1181 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.ConfiguredTaskAwaitable1.ConfiguredTaskAwaiter.GetResult()
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<>c.<<GetStateAsync>b__62_0>d.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Amqp/AmqpReceiver.cs:line 1164
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Threading.Tasks.ValueTask1.get_Result() at Azure.Messaging.ServiceBus.ServiceBusRetryPolicy.<RunOperation>d__212.MoveNext() in //sdk/servicebus/Azure.Messaging.ServiceBus/src/Primitives/ServiceBusRetryPolicy.cs:line 144
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.ServiceBusRetryPolicy.<RunOperation>d__212.MoveNext() in /_/sdk/servicebus/Azure.Messaging.ServiceBus/src/Primitives/ServiceBusRetryPolicy.cs:line 164 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Threading.Tasks.ValueTask1.get_Result()
> at Azure.Messaging.ServiceBus.Amqp.AmqpReceiver.<GetStateAsync>d__62.MoveNext
>
> Stack trace
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 176
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 187
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 187
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at Azure.Messaging.ServiceBus.Tests.Processor.SessionProcessorLiveTests.<ProcessSessionMessage>d__2.MoveNext() in D:\a\_work\1\s\sdk\servicebus\Azure.Messaging.ServiceBus\tests\Processor\SessionProcessorLiveTests.cs:line 188
> --- End of stack trace from previous location where exception was thrown ---
> at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
> at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
> at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
> at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted()
> at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter)
> at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke)
> at NUnit.Framework.Internal.Commands.TestMethodCommand.RunTestMethod(TestExecutionContext context)
> at NUnit.Framework.Internal.Commands.TestMethodCommand.Execute(TestExecutionContext context)
> at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.<>c__DisplayClass1_0.<Execute>b__0()
> at NUnit.Framework.Internal.Commands.BeforeAndAfterTestCommand.RunTestMethodInThreadAbortSafeZone(TestExecutionContext context, Action action)
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1034081&view=results
|
non_process
|
test failing in nightly runs servicebus nightly test runs are failing with error message multiple failures or warnings in test azure messaging servicebus servicebusexception the session lock was lost request a new session receiver sessionlocklost azure messaging servicebus servicebusexception the write operation failed see inner exception errorcode connectionreset servicecommunicationproblem system io ioexception the write operation failed see inner exception system net sockets socketexception an existing connection was forcibly closed by the remote host at microsoft azure amqp transport transportstream endwrite iasyncresult asyncresult at system net security sslstream startwriting byte buffer offset count asyncprotocolrequest asyncrequest at system net security sslstream processwrite byte buffer offset count asyncprotocolrequest asyncrequest end of inner exception stack trace at system net security sslstream processwrite byte buffer offset count asyncprotocolrequest asyncrequest at system net security sslstream beginwrite byte buffer offset count asynccallback asynccallback object asyncstate at microsoft azure amqp transport tlstransport writeasync transportasynccallbackargs args at microsoft azure amqp asyncio asyncwriter writecore end of inner exception stack trace end of inner exception stack trace at azure messaging servicebus amqp amqpreceiver throwifsessionlocklost in sdk servicebus azure messaging servicebus src amqp amqpreceiver cs line at azure messaging servicebus amqp amqpreceiver d movenext in sdk servicebus azure messaging servicebus src amqp amqpreceiver cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices configuredtaskawaiter getresult at azure messaging servicebus amqp amqpreceiver d movenext in sdk servicebus azure messaging servicebus src amqp amqpreceiver cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices configuredtaskawaiter getresult at azure messaging servicebus amqp amqpreceiver c b d movenext in sdk servicebus azure messaging servicebus src amqp amqpreceiver cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system threading tasks get result at azure messaging servicebus servicebusretrypolicy d movenext in sdk servicebus azure messaging servicebus src primitives servicebusretrypolicy cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at azure messaging servicebus servicebusretrypolicy d movenext in sdk servicebus azure messaging servicebus src primitives servicebusretrypolicy cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system threading tasks get result at azure messaging servicebus amqp amqpreceiver d movenext stack trace at azure messaging servicebus tests processor sessionprocessorlivetests d movenext in d a work s sdk servicebus azure messaging servicebus tests processor sessionprocessorlivetests cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at azure messaging servicebus tests processor sessionprocessorlivetests d movenext in d a work s sdk servicebus azure messaging servicebus tests processor sessionprocessorlivetests cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at azure messaging servicebus tests processor sessionprocessorlivetests d movenext in d a work s sdk servicebus azure messaging servicebus tests processor sessionprocessorlivetests cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at azure messaging servicebus tests processor sessionprocessorlivetests d movenext in d a work s sdk servicebus azure messaging servicebus tests processor sessionprocessorlivetests cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at nunit framework internal taskawaitadapter genericadapter blockuntilcompleted at nunit framework internal messagepumpstrategy nomessagepumpstrategy waitforcompletion awaitadapter awaiter at nunit framework internal asynctosyncadapter await func invoke at nunit framework internal commands testmethodcommand runtestmethod testexecutioncontext context at nunit framework internal commands testmethodcommand execute testexecutioncontext context at nunit framework internal commands beforeandaftertestcommand c b at nunit framework internal commands beforeandaftertestcommand runtestmethodinthreadabortsafezone testexecutioncontext context action action for more details check here
| 0
|
409,901
| 27,757,117,114
|
IssuesEvent
|
2023-03-16 04:06:09
|
42organization/42gg.client
|
https://api.github.com/repos/42organization/42gg.client
|
closed
|
[Document] 와이어프레임 제작
|
documentation
|
## DESC
1. 관리자 페이지 view
2. 상점 페이지 view
3. (3기) 사이트 renewal view
## TODO
- [x] #639
- [ ] 상점 페이지
- [ ] (renewal) 메인 페이지
## ETC
- 디자이너가 2월 초까지만 풀타임이 가능하기에 왠만하면 그전에 기능이 fix되야함
|
1.0
|
[Document] 와이어프레임 제작 - ## DESC
1. 관리자 페이지 view
2. 상점 페이지 view
3. (3기) 사이트 renewal view
## TODO
- [x] #639
- [ ] 상점 페이지
- [ ] (renewal) 메인 페이지
## ETC
- 디자이너가 2월 초까지만 풀타임이 가능하기에 왠만하면 그전에 기능이 fix되야함
|
non_process
|
와이어프레임 제작 desc 관리자 페이지 view 상점 페이지 view 사이트 renewal view todo 상점 페이지 renewal 메인 페이지 etc 디자이너가 초까지만 풀타임이 가능하기에 왠만하면 그전에 기능이 fix되야함
| 0
|
56,222
| 3,078,599,914
|
IssuesEvent
|
2015-08-21 11:25:14
|
pavel-pimenov/flylinkdc-r5xx
|
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
|
closed
|
Отказ от использования WinRAR
|
bug imported Priority-Medium
|
_From [birkoff.anarchist](https://code.google.com/u/birkoff.anarchist/) on June 22, 2011 11:33:06_
На вики-странице посвященной компиляции в качестве возможных проблем указывается отсутствия архиватора winrar. Если просмотреть все bat-файлы, то unrar используется только в
FlyLinkDC\compiled\Settings\common\get_geo_ip.bat
С учетом того, что параллельно в полную силу используется архиватор 7zip, в сбоку которого уже входит модуль unrar, в приведенном файле необходимо заменить
unrar.exe x customlocations-5.rar -y
на
7z x customlocations-5.rar -y
И необходимость в установке платного WinRAR полностью отпадет.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=500_
|
1.0
|
Отказ от использования WinRAR - _From [birkoff.anarchist](https://code.google.com/u/birkoff.anarchist/) on June 22, 2011 11:33:06_
На вики-странице посвященной компиляции в качестве возможных проблем указывается отсутствия архиватора winrar. Если просмотреть все bat-файлы, то unrar используется только в
FlyLinkDC\compiled\Settings\common\get_geo_ip.bat
С учетом того, что параллельно в полную силу используется архиватор 7zip, в сбоку которого уже входит модуль unrar, в приведенном файле необходимо заменить
unrar.exe x customlocations-5.rar -y
на
7z x customlocations-5.rar -y
И необходимость в установке платного WinRAR полностью отпадет.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=500_
|
non_process
|
отказ от использования winrar from on june на вики странице посвященной компиляции в качестве возможных проблем указывается отсутствия архиватора winrar если просмотреть все bat файлы то unrar используется только в flylinkdc compiled settings common get geo ip bat с учетом того что параллельно в полную силу используется архиватор в сбоку которого уже входит модуль unrar в приведенном файле необходимо заменить unrar exe x customlocations rar y на x customlocations rar y и необходимость в установке платного winrar полностью отпадет original issue
| 0
|
296,939
| 9,134,975,689
|
IssuesEvent
|
2019-02-26 02:07:32
|
python/mypy
|
https://api.github.com/repos/python/mypy
|
closed
|
Union of different-length Tuple types should agree for corresponding items
|
bug priority-1-normal topic-union-types
|
See https://github.com/python/typeshed/pull/721#discussion_r90098017. The issue is that after
```python
T = Union[Tuple[int, int], Tuple[int, int, str]]
def f() -> T: pass
a = f()
reveal_type(a[0])
reveal_type(a[2])
```
the item types are considered to be Any. A better approach would be to understand that a[0] is always an int while a[2] is invalid without some check (however we can't do isinstance() checks on tuple lengths so that's problematic).
|
1.0
|
Union of different-length Tuple types should agree for corresponding items - See https://github.com/python/typeshed/pull/721#discussion_r90098017. The issue is that after
```python
T = Union[Tuple[int, int], Tuple[int, int, str]]
def f() -> T: pass
a = f()
reveal_type(a[0])
reveal_type(a[2])
```
the item types are considered to be Any. A better approach would be to understand that a[0] is always an int while a[2] is invalid without some check (however we can't do isinstance() checks on tuple lengths so that's problematic).
|
non_process
|
union of different length tuple types should agree for corresponding items see the issue is that after python t union tuple def f t pass a f reveal type a reveal type a the item types are considered to be any a better approach would be to understand that a is always an int while a is invalid without some check however we can t do isinstance checks on tuple lengths so that s problematic
| 0
|
313,250
| 9,559,023,239
|
IssuesEvent
|
2019-05-03 15:35:11
|
canonical-web-and-design/www.ubuntu.com
|
https://api.github.com/repos/canonical-web-and-design/www.ubuntu.com
|
closed
|
Update release chart on /server page
|
Priority: High
|
The release chart doesn't have the ESM support additions to match the updated release cycle page - https://www.ubuntu.com/about/release-cycle
---
*Reported from: https://www.ubuntu.com/server*
|
1.0
|
Update release chart on /server page - The release chart doesn't have the ESM support additions to match the updated release cycle page - https://www.ubuntu.com/about/release-cycle
---
*Reported from: https://www.ubuntu.com/server*
|
non_process
|
update release chart on server page the release chart doesn t have the esm support additions to match the updated release cycle page reported from
| 0
|
164,336
| 25,951,239,396
|
IssuesEvent
|
2022-12-17 16:27:40
|
ita-social-projects/horondi_client_fe
|
https://api.github.com/repos/ita-social-projects/horondi_client_fe
|
closed
|
Redesign Checkout component according to the new design
|
FrontEnd part redesign
|
Redesign Checkout component.
Expected result:

|
1.0
|
Redesign Checkout component according to the new design - Redesign Checkout component.
Expected result:

|
non_process
|
redesign checkout component according to the new design redesign checkout component expected result
| 0
|
4,127
| 7,086,111,340
|
IssuesEvent
|
2018-01-11 13:32:28
|
bazelbuild/continuous-integration
|
https://api.github.com/repos/bazelbuild/continuous-integration
|
closed
|
Increase performance of CI builds
|
P1 process
|
So far a global tests takes around 1h30 minutes, most of it is taken from downstream jobs, the top 3 slowers are:
1. [TensorFlow](http://ci.bazel.io/job/Global/job/TensorFlow/workflow-stage/): ~1h15, capped by windows build (linux build is below 20 minutes)
2. [bazel-docker-tests](http://ci.bazel.io/job/Global/job/bazel-docker-tests/workflow-stage/): ~1h. 2 issues: all the tests are exclusive and non incremental (any change to bazel distribution artifact re-run all the tests)
3. [bazel-tests](http://ci.bazel.io/job/Global/job/bazel-tests/workflow-stage/): ~35min. Tests are 30minutes on all config, mostly capped by integration tests.
Project using TensorFlow are next on the lists.
Proposals:
1. Distributed caching (#97) should help a lot for 1 and 3 and all the other.
2. To reduce the whole pipeline, increasing our capacity should help: add 4 mac server and 4 windows to have all the fleet with 8 slaves of each. It should also decrease the pressure on CI during daytime in Europe when people use a lot CI.
3. For 1. we must make TF move out of the ugly CI Windows setup. /cc @meteorcloudy. At least we could trick a bit to reuse the local cache.
4. To reduce the cap time on 3, we could make integration tests not depends on packed bazel but instead on each tools directly, that would remove the ZIP phase from the tests but also we could make tests depends on only part of the tools instead. It would however increase the maintenance burden of those tests.
5. bazel-docker-tests time can only be reduced by sharding the tests among several workers, apart from using remote execution, the only way would be to create more jobs for now and manually shard those tests.
I believe we can easily goes to a Global test under 30 minutes if we do all that.
/cc @buchgr @aehlig @ulfjack FYI
|
1.0
|
Increase performance of CI builds - So far a global tests takes around 1h30 minutes, most of it is taken from downstream jobs, the top 3 slowers are:
1. [TensorFlow](http://ci.bazel.io/job/Global/job/TensorFlow/workflow-stage/): ~1h15, capped by windows build (linux build is below 20 minutes)
2. [bazel-docker-tests](http://ci.bazel.io/job/Global/job/bazel-docker-tests/workflow-stage/): ~1h. 2 issues: all the tests are exclusive and non incremental (any change to bazel distribution artifact re-run all the tests)
3. [bazel-tests](http://ci.bazel.io/job/Global/job/bazel-tests/workflow-stage/): ~35min. Tests are 30minutes on all config, mostly capped by integration tests.
Project using TensorFlow are next on the lists.
Proposals:
1. Distributed caching (#97) should help a lot for 1 and 3 and all the other.
2. To reduce the whole pipeline, increasing our capacity should help: add 4 mac server and 4 windows to have all the fleet with 8 slaves of each. It should also decrease the pressure on CI during daytime in Europe when people use a lot CI.
3. For 1. we must make TF move out of the ugly CI Windows setup. /cc @meteorcloudy. At least we could trick a bit to reuse the local cache.
4. To reduce the cap time on 3, we could make integration tests not depends on packed bazel but instead on each tools directly, that would remove the ZIP phase from the tests but also we could make tests depends on only part of the tools instead. It would however increase the maintenance burden of those tests.
5. bazel-docker-tests time can only be reduced by sharding the tests among several workers, apart from using remote execution, the only way would be to create more jobs for now and manually shard those tests.
I believe we can easily goes to a Global test under 30 minutes if we do all that.
/cc @buchgr @aehlig @ulfjack FYI
|
process
|
increase performance of ci builds so far a global tests takes around minutes most of it is taken from downstream jobs the top slowers are capped by windows build linux build is below minutes issues all the tests are exclusive and non incremental any change to bazel distribution artifact re run all the tests tests are on all config mostly capped by integration tests project using tensorflow are next on the lists proposals distributed caching should help a lot for and and all the other to reduce the whole pipeline increasing our capacity should help add mac server and windows to have all the fleet with slaves of each it should also decrease the pressure on ci during daytime in europe when people use a lot ci for we must make tf move out of the ugly ci windows setup cc meteorcloudy at least we could trick a bit to reuse the local cache to reduce the cap time on we could make integration tests not depends on packed bazel but instead on each tools directly that would remove the zip phase from the tests but also we could make tests depends on only part of the tools instead it would however increase the maintenance burden of those tests bazel docker tests time can only be reduced by sharding the tests among several workers apart from using remote execution the only way would be to create more jobs for now and manually shard those tests i believe we can easily goes to a global test under minutes if we do all that cc buchgr aehlig ulfjack fyi
| 1
|
21,504
| 29,670,495,651
|
IssuesEvent
|
2023-06-11 10:49:06
|
giorgiopizz/mkShapesRDF
|
https://api.github.com/repos/giorgiopizz/mkShapesRDF
|
closed
|
[PROC] Verify mRDF Vary function with Snapshot
|
processor
|
With [this commit](https://github.com/giorgiopizz/mkShapesRDF/commit/63a1d0be7fbf3c78fae737810f7e6aa1b5d8ed77) it was added support to the classic Vary function to mRDF.
The problem was that bare RDataFrames do not support the Vary functionality with Snaphot (for reference see [this](https://github.com/root-project/root/issues/11218)).
By parsing c++ definition strings (expression passed in `mRDF.Define`) it should be possible to handle variations.
It's important to note that one should first define variations, e.g. JES, and then compute kinematic variables that rely on varied quantities: `mjj` relies on `CleanJet_pt` which is varied by JES, so the workflow should be:
1. Define nominal (`CleanJet_pt`)
2. Define variations (JES)
3. Define other variables (`mjj`)
This functionality should be tested!
|
1.0
|
[PROC] Verify mRDF Vary function with Snapshot - With [this commit](https://github.com/giorgiopizz/mkShapesRDF/commit/63a1d0be7fbf3c78fae737810f7e6aa1b5d8ed77) it was added support to the classic Vary function to mRDF.
The problem was that bare RDataFrames do not support the Vary functionality with Snaphot (for reference see [this](https://github.com/root-project/root/issues/11218)).
By parsing c++ definition strings (expression passed in `mRDF.Define`) it should be possible to handle variations.
It's important to note that one should first define variations, e.g. JES, and then compute kinematic variables that rely on varied quantities: `mjj` relies on `CleanJet_pt` which is varied by JES, so the workflow should be:
1. Define nominal (`CleanJet_pt`)
2. Define variations (JES)
3. Define other variables (`mjj`)
This functionality should be tested!
|
process
|
verify mrdf vary function with snapshot with it was added support to the classic vary function to mrdf the problem was that bare rdataframes do not support the vary functionality with snaphot for reference see by parsing c definition strings expression passed in mrdf define it should be possible to handle variations it s important to note that one should first define variations e g jes and then compute kinematic variables that rely on varied quantities mjj relies on cleanjet pt which is varied by jes so the workflow should be define nominal cleanjet pt define variations jes define other variables mjj this functionality should be tested
| 1
|
127,538
| 17,294,787,329
|
IssuesEvent
|
2021-07-25 13:59:33
|
mexyn/statev_v2_issues
|
https://api.github.com/repos/mexyn/statev_v2_issues
|
closed
|
Cop Attachments bei Frauen
|
gamedesign solved
|
<!-- Bitte die Vorlage unten vollständig ausfüllen -->
**Character Name**
Amber Reaper & diverse andere
**Auftrittszeitpunkt (Datum / Uhrzeit)**
16.07.2021
**Beobachtetes Verhalten**
Amber Reaper hat Cop Kleidung an (Handschellen & Marke)
Sie selbe kann diese Gegenstände an sich nicht sehen, allerdings alle anderen.
Dies tritt bei diversen Frauen auf, allerdings nicht bei allen.
**Erwartetes Verhalten**
Marke und Handschellen nur im Dienst bei Cops.
**Schritte um den Fehler nachvollziehen zu können**
<!--- Beschreibe Schritt für Schritt wie man den Fehler nachstellen kann -->
**Monitorauflösung (nur wenn Falsche Darstellung in der UI)**
<!--- Beschreibe Schritt für Schritt wie man den Fehler nachstellen kann -->
**Optional: Video / Bilder des Fehlers**



|
1.0
|
Cop Attachments bei Frauen - <!-- Bitte die Vorlage unten vollständig ausfüllen -->
**Character Name**
Amber Reaper & diverse andere
**Auftrittszeitpunkt (Datum / Uhrzeit)**
16.07.2021
**Beobachtetes Verhalten**
Amber Reaper hat Cop Kleidung an (Handschellen & Marke)
Sie selbe kann diese Gegenstände an sich nicht sehen, allerdings alle anderen.
Dies tritt bei diversen Frauen auf, allerdings nicht bei allen.
**Erwartetes Verhalten**
Marke und Handschellen nur im Dienst bei Cops.
**Schritte um den Fehler nachvollziehen zu können**
<!--- Beschreibe Schritt für Schritt wie man den Fehler nachstellen kann -->
**Monitorauflösung (nur wenn Falsche Darstellung in der UI)**
<!--- Beschreibe Schritt für Schritt wie man den Fehler nachstellen kann -->
**Optional: Video / Bilder des Fehlers**



|
non_process
|
cop attachments bei frauen character name amber reaper diverse andere auftrittszeitpunkt datum uhrzeit beobachtetes verhalten amber reaper hat cop kleidung an handschellen marke sie selbe kann diese gegenstände an sich nicht sehen allerdings alle anderen dies tritt bei diversen frauen auf allerdings nicht bei allen erwartetes verhalten marke und handschellen nur im dienst bei cops schritte um den fehler nachvollziehen zu können monitorauflösung nur wenn falsche darstellung in der ui optional video bilder des fehlers
| 0
|
704,410
| 24,196,221,018
|
IssuesEvent
|
2022-09-24 00:19:39
|
openservicemesh/osm
|
https://api.github.com/repos/openservicemesh/osm
|
closed
|
ServiceMonitor feature
|
size/S priority/P3 kind/feature-request stale
|
**Please describe the Improvement and/or Feature Request**
<!-- A clear and concise description of the proposal. -->
**Scope (please mark with X where applicable)**
- New Functionality [ ]
- Install [ ]
- SMI Traffic Access Policy [ ]
- SMI Traffic Specs Policy [ ]
- SMI Traffic Split Policy [ ]
- Permissive Traffic Policy [ ]
- Ingress [ ]
- Egress [ ]
- Envoy Control Plane [ ]
- CLI Tool [ ]
- Metrics [ ]
- Certificate Management [ ]
- Sidecar Injection [ ]
- Logging [ ]
- Debugging [ ]
- Tests [ ]
- CI System [ ]
- Demo [ ]
- Project Release [ ]
**Possible use cases**
<!-- A clear and concise description of possible use cases for this proposal. -->
A lot of Helm charts could create Kubernetes resources (ServiceMonitor, PodMonitor, ...) managed by the [Prometheus Operator](https://prometheus-operator.dev/).
Ex:
- [keda](https://github.com/kedacore/charts/blob/main/keda/templates/26-metrics-podmonitor.yaml)
- [grafana](https://github.com/grafana/helm-charts/blob/main/charts/grafana/templates/servicemonitor.yaml)
It would be interesting to have such a feature into the Helm chart.
|
1.0
|
ServiceMonitor feature - **Please describe the Improvement and/or Feature Request**
<!-- A clear and concise description of the proposal. -->
**Scope (please mark with X where applicable)**
- New Functionality [ ]
- Install [ ]
- SMI Traffic Access Policy [ ]
- SMI Traffic Specs Policy [ ]
- SMI Traffic Split Policy [ ]
- Permissive Traffic Policy [ ]
- Ingress [ ]
- Egress [ ]
- Envoy Control Plane [ ]
- CLI Tool [ ]
- Metrics [ ]
- Certificate Management [ ]
- Sidecar Injection [ ]
- Logging [ ]
- Debugging [ ]
- Tests [ ]
- CI System [ ]
- Demo [ ]
- Project Release [ ]
**Possible use cases**
<!-- A clear and concise description of possible use cases for this proposal. -->
A lot of Helm charts could create Kubernetes resources (ServiceMonitor, PodMonitor, ...) managed by the [Prometheus Operator](https://prometheus-operator.dev/).
Ex:
- [keda](https://github.com/kedacore/charts/blob/main/keda/templates/26-metrics-podmonitor.yaml)
- [grafana](https://github.com/grafana/helm-charts/blob/main/charts/grafana/templates/servicemonitor.yaml)
It would be interesting to have such a feature into the Helm chart.
|
non_process
|
servicemonitor feature please describe the improvement and or feature request scope please mark with x where applicable new functionality install smi traffic access policy smi traffic specs policy smi traffic split policy permissive traffic policy ingress egress envoy control plane cli tool metrics certificate management sidecar injection logging debugging tests ci system demo project release possible use cases a lot of helm charts could create kubernetes resources servicemonitor podmonitor managed by the ex it would be interesting to have such a feature into the helm chart
| 0
|
17,077
| 22,578,895,875
|
IssuesEvent
|
2022-06-28 09:47:22
|
alphagov/govuk-design-system
|
https://api.github.com/repos/alphagov/govuk-design-system
|
closed
|
Update branding and create merchandise / swag for Design System Day 2022
|
🕔 weeks process shared ownership
|
## What
Update branding from Design System Day 2021, and create merch for this year's event.
## Why
So we can give attendees something following the event.
## Who needs to work on this
TBC
## Done when
- [x] Updated branding
- [x] Agreed other swag
|
1.0
|
Update branding and create merchandise / swag for Design System Day 2022 - ## What
Update branding from Design System Day 2021, and create merch for this year's event.
## Why
So we can give attendees something following the event.
## Who needs to work on this
TBC
## Done when
- [x] Updated branding
- [x] Agreed other swag
|
process
|
update branding and create merchandise swag for design system day what update branding from design system day and create merch for this year s event why so we can give attendees something following the event who needs to work on this tbc done when updated branding agreed other swag
| 1
|
156,775
| 24,625,660,019
|
IssuesEvent
|
2022-10-16 13:39:25
|
dotnet/efcore
|
https://api.github.com/repos/dotnet/efcore
|
closed
|
ParameterExtractingExpressionVisitor::GetValue
|
closed-by-design customer-reported
|
EFCore: master
Could you change
**private object GetValue(Expression expression, out string parameterName)**
to
**protected virtual object GetValue(Expression expression, out string parameterName)**
?
My Data Provider for EF needs in overriding of this method for preparing (transforming) expression before local evaluation.
Thanks.
|
1.0
|
ParameterExtractingExpressionVisitor::GetValue - EFCore: master
Could you change
**private object GetValue(Expression expression, out string parameterName)**
to
**protected virtual object GetValue(Expression expression, out string parameterName)**
?
My Data Provider for EF needs in overriding of this method for preparing (transforming) expression before local evaluation.
Thanks.
|
non_process
|
parameterextractingexpressionvisitor getvalue efcore master could you change private object getvalue expression expression out string parametername to protected virtual object getvalue expression expression out string parametername my data provider for ef needs in overriding of this method for preparing transforming expression before local evaluation thanks
| 0
|
123,150
| 16,448,989,108
|
IssuesEvent
|
2021-05-21 00:47:09
|
revelrylabs/harmonium
|
https://api.github.com/repos/revelrylabs/harmonium
|
reopened
|
As a User I want simple default paragraph styles
|
designy stale
|
**I'm submitting a ...**
- [x] feature request
<!-- feel free to remove any items that don't apply to this issue -->
**What is the current behavior?**
All `<p>` tags have a default attribute: `max-width: 70rem`

A a user this is somewhat baffling. In its current state there needs to be a removal/override on `<p>` tags. While this is fine for Harmonium documentation, this should not be the default for the framework.
### Scenario 1 - as a Harmonium user
- [ ] When I add a p tag
- [ ] Then it does not have a max-width attribute
**For bugs: please provide steps to reproduce and tell us a little about your environment (e.g., what OS and version of the package you're using)**
_give us enough information to properly diagnose the bug_
**What is the expected behavior?**
_tell us what you think should happen instead of what's happening now_
**For feature requests: why should the behavior change?**
_explain why your feature request is a good idea_
**Other information (we love stack traces!)**
<!-- every little bit helps -->
|
1.0
|
As a User I want simple default paragraph styles - **I'm submitting a ...**
- [x] feature request
<!-- feel free to remove any items that don't apply to this issue -->
**What is the current behavior?**
All `<p>` tags have a default attribute: `max-width: 70rem`

A a user this is somewhat baffling. In its current state there needs to be a removal/override on `<p>` tags. While this is fine for Harmonium documentation, this should not be the default for the framework.
### Scenario 1 - as a Harmonium user
- [ ] When I add a p tag
- [ ] Then it does not have a max-width attribute
**For bugs: please provide steps to reproduce and tell us a little about your environment (e.g., what OS and version of the package you're using)**
_give us enough information to properly diagnose the bug_
**What is the expected behavior?**
_tell us what you think should happen instead of what's happening now_
**For feature requests: why should the behavior change?**
_explain why your feature request is a good idea_
**Other information (we love stack traces!)**
<!-- every little bit helps -->
|
non_process
|
as a user i want simple default paragraph styles i m submitting a feature request what is the current behavior all tags have a default attribute max width a a user this is somewhat baffling in its current state there needs to be a removal override on tags while this is fine for harmonium documentation this should not be the default for the framework scenario as a harmonium user when i add a p tag then it does not have a max width attribute for bugs please provide steps to reproduce and tell us a little about your environment e g what os and version of the package you re using give us enough information to properly diagnose the bug what is the expected behavior tell us what you think should happen instead of what s happening now for feature requests why should the behavior change explain why your feature request is a good idea other information we love stack traces
| 0
|
1,213
| 3,715,967,337
|
IssuesEvent
|
2016-03-03 05:31:02
|
shakeelmohamed/Today-Was
|
https://api.github.com/repos/shakeelmohamed/Today-Was
|
closed
|
Travis CI - auto deployment + testing
|
enhancement process
|
I already did everything with egress.
Docs: http://about.travis-ci.org/docs/user/deployment/heroku/
- [ ] Add a `beta` branch, which contains code form `develop` that has passed all tests on Travis CI. Be sure to document this on the wiki somewhere.
- [ ] Write some non-db dependent tests
- [x] Configure[ databases for travis CI](http://about.travis-ci.org/docs/user/database-setup/#multiple-database-systems)
- [ ] Figure out command line scripts to give to Travis CI
- [ ] Some optimization tips: https://coderwall.com/p/x8exja
|
1.0
|
Travis CI - auto deployment + testing - I already did everything with egress.
Docs: http://about.travis-ci.org/docs/user/deployment/heroku/
- [ ] Add a `beta` branch, which contains code form `develop` that has passed all tests on Travis CI. Be sure to document this on the wiki somewhere.
- [ ] Write some non-db dependent tests
- [x] Configure[ databases for travis CI](http://about.travis-ci.org/docs/user/database-setup/#multiple-database-systems)
- [ ] Figure out command line scripts to give to Travis CI
- [ ] Some optimization tips: https://coderwall.com/p/x8exja
|
process
|
travis ci auto deployment testing i already did everything with egress docs add a beta branch which contains code form develop that has passed all tests on travis ci be sure to document this on the wiki somewhere write some non db dependent tests configure figure out command line scripts to give to travis ci some optimization tips
| 1
|
205,053
| 23,296,039,019
|
IssuesEvent
|
2022-08-06 15:33:50
|
turkdevops/grafana
|
https://api.github.com/repos/turkdevops/grafana
|
closed
|
WS-2022-0008 (Medium) detected in node-forge-0.9.0.tgz - autoclosed
|
security vulnerability
|
## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p>
<p>Found in base branch: <b>datasource-meta</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2022-0008 (Medium) detected in node-forge-0.9.0.tgz - autoclosed - ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.2.1.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p>
<p>Found in base branch: <b>datasource-meta</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in node forge tgz autoclosed ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in head commit a href found in base branch datasource meta vulnerability details the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution webpack dev server step up your open source security game with mend
| 0
|
858
| 3,317,235,079
|
IssuesEvent
|
2015-11-06 20:38:45
|
broadinstitute/gatk
|
https://api.github.com/repos/broadinstitute/gatk
|
closed
|
Create mapping from Human readable names to Ids for Google Genomics Reference API
|
DataflowPreprocessingPipeline Spark
|
Right now, the "names" in the table mapping (in RefAPISource) are things like EOSt9JOVhp3jkwE. What we want is "GRCh37".
|
1.0
|
Create mapping from Human readable names to Ids for Google Genomics Reference API - Right now, the "names" in the table mapping (in RefAPISource) are things like EOSt9JOVhp3jkwE. What we want is "GRCh37".
|
process
|
create mapping from human readable names to ids for google genomics reference api right now the names in the table mapping in refapisource are things like what we want is
| 1
|
8,142
| 11,352,110,724
|
IssuesEvent
|
2020-01-24 12:54:00
|
peopledoc/procrastinate
|
https://api.github.com/repos/peopledoc/procrastinate
|
opened
|
DeprecationWarning with Python 3.8
|
Type: Process
|
While running the tests on Python 3.8 we get a `DeprecationWarning`:
```
DeprecationWarning: The loop argument is deprecated since Python 3.8, and scheduled for removal
in Python 3.10.
```
We get this warning several times.
|
1.0
|
DeprecationWarning with Python 3.8 - While running the tests on Python 3.8 we get a `DeprecationWarning`:
```
DeprecationWarning: The loop argument is deprecated since Python 3.8, and scheduled for removal
in Python 3.10.
```
We get this warning several times.
|
process
|
deprecationwarning with python while running the tests on python we get a deprecationwarning deprecationwarning the loop argument is deprecated since python and scheduled for removal in python we get this warning several times
| 1
|
22,205
| 30,758,659,656
|
IssuesEvent
|
2023-07-29 12:02:38
|
DxytJuly3/gitalk_blog
|
https://api.github.com/repos/DxytJuly3/gitalk_blog
|
opened
|
[Linux] 一篇文章, 掌握Linux进程信号: 信号的产生、深入理解信号的处理与捕捉、信号在内核中的表示、进程的内核态与用户态转换分析、volatile关键字... - July.cc Blogs
|
Gitalk /posts/Linux-processSignals
|
https://www.julysblog.cn/posts/Linux-processSignals
进程信号, 在Linux系统的学习中, 是一个非常重要的概念. 我们可以通过向进程发送信号来让进程执行某些指定的动作.
|
1.0
|
[Linux] 一篇文章, 掌握Linux进程信号: 信号的产生、深入理解信号的处理与捕捉、信号在内核中的表示、进程的内核态与用户态转换分析、volatile关键字... - July.cc Blogs - https://www.julysblog.cn/posts/Linux-processSignals
进程信号, 在Linux系统的学习中, 是一个非常重要的概念. 我们可以通过向进程发送信号来让进程执行某些指定的动作.
|
process
|
一篇文章 掌握linux进程信号 信号的产生、深入理解信号的处理与捕捉、信号在内核中的表示、进程的内核态与用户态转换分析、volatile关键字 july cc blogs 进程信号 在linux系统的学习中 是一个非常重要的概念 我们可以通过向进程发送信号来让进程执行某些指定的动作
| 1
|
537,308
| 15,726,922,134
|
IssuesEvent
|
2021-03-29 12:00:14
|
protofire/omen-exchange
|
https://api.github.com/repos/protofire/omen-exchange
|
opened
|
After switching account/network, pending modal is still showing
|
priority:low
|
<img width="1329" alt="Bildschirmfoto 2021-03-29 um 13 59 37" src="https://user-images.githubusercontent.com/5337809/112833459-0e575d00-9097-11eb-8134-8d2864caab3d.png">
|
1.0
|
After switching account/network, pending modal is still showing - <img width="1329" alt="Bildschirmfoto 2021-03-29 um 13 59 37" src="https://user-images.githubusercontent.com/5337809/112833459-0e575d00-9097-11eb-8134-8d2864caab3d.png">
|
non_process
|
after switching account network pending modal is still showing img width alt bildschirmfoto um src
| 0
|
20,089
| 26,602,330,294
|
IssuesEvent
|
2023-01-23 16:36:44
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/attributes] resource attributes sent to incorrect tenant
|
bug priority:p2 processor/attributes never stale
|
**Describe the bug**
multi-tenant use cases in the Loki exporter were introduced through https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/12415, which allows setting and creating tenantIDs from resource attribute values. This is awesome and is the exact functionality we are looking for, but we are somehow seeing attribute values being sent to the wrong tenant, especially when logging at scale.
For example, if we configure the loki exporter to set the tenantID from the "namespace" attribute, we are seeing many logs from namespace `foo` show up in the namespace `bar` tenant in Loki.
**Steps to reproduce**
This has been occurring using the opentelemetry collector helm chart with the following configuration. We are following the recommendations in the documentation to use the groupbyattrs processor, as well as batching.
```yaml
config:
receivers:
fluentforward:
endpoint: 0.0.0.0:8006
processors:
groupbyattrs:
keys:
- namespace_name
filter/exclude:
logs:
exclude:
match_type: regexp
resource_attributes:
- Key: namespace_name
Value: istio-system|kube-system
exporters:
loki:
endpoint: http://loki-distributed-gateway.loki.svc.cluster.local/loki/api/v1/push
format: json
tenant:
source: attributes
value: namespace_name
labels:
resource:
namespace_name: ""
attributes:
stream: ""
app: ""
pod_name: ""
account: ""
cluster_name: ""
region: ""
container_name: ""
service:
pipelines:
logs:
exporters: [loki]
processors: [memory_limiter, batch, groupbyattrs, filter/exclude]
receivers: [fluentforward]
```
To reproduce, we noticed this occur more frequently during spiky loads at a higher volume, but also noticed it occur sometimes when the opentelemetry collector was killed or rescheduled onto a new node.
As part of troubleshooting we have also tested multiple combinations of this config, such as batching and grouping after the filter, as well as before, and also not batching at all. None of these have resulted in proper tenants.
**What did you expect to see?**
We expect to see all logs from namespace `foo` only be in tenant `foo` and all logs from namespace `bar` only be in tenant `bar`.
**What did you see instead?**
Logs from every namespace appear in other tenants.
below are some screenshots from Grafana where we expect to only see logs from the `foo` namespace.

Instead, we see many logs from other namespaces as well (even though our X-Scope-OrgID is set to the `foo` tenant)

**What version did you use?**
Version: v0.57.2
**Additional context**
We noticed in testing environments that with small non-spiky loads the logs were grouped and exported to the correct tenant, but in high-volume spiky environments the logs were not.
|
1.0
|
[processor/attributes] resource attributes sent to incorrect tenant - **Describe the bug**
multi-tenant use cases in the Loki exporter were introduced through https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/12415, which allows setting and creating tenantIDs from resource attribute values. This is awesome and is the exact functionality we are looking for, but we are somehow seeing attribute values being sent to the wrong tenant, especially when logging at scale.
For example, if we configure the loki exporter to set the tenantID from the "namespace" attribute, we are seeing many logs from namespace `foo` show up in the namespace `bar` tenant in Loki.
**Steps to reproduce**
This has been occurring using the opentelemetry collector helm chart with the following configuration. We are following the recommendations in the documentation to use the groupbyattrs processor, as well as batching.
```yaml
config:
receivers:
fluentforward:
endpoint: 0.0.0.0:8006
processors:
groupbyattrs:
keys:
- namespace_name
filter/exclude:
logs:
exclude:
match_type: regexp
resource_attributes:
- Key: namespace_name
Value: istio-system|kube-system
exporters:
loki:
endpoint: http://loki-distributed-gateway.loki.svc.cluster.local/loki/api/v1/push
format: json
tenant:
source: attributes
value: namespace_name
labels:
resource:
namespace_name: ""
attributes:
stream: ""
app: ""
pod_name: ""
account: ""
cluster_name: ""
region: ""
container_name: ""
service:
pipelines:
logs:
exporters: [loki]
processors: [memory_limiter, batch, groupbyattrs, filter/exclude]
receivers: [fluentforward]
```
To reproduce, we noticed this occur more frequently during spiky loads at a higher volume, but also noticed it occur sometimes when the opentelemetry collector was killed or rescheduled onto a new node.
As part of troubleshooting we have also tested multiple combinations of this config, such as batching and grouping after the filter, as well as before, and also not batching at all. None of these have resulted in proper tenants.
**What did you expect to see?**
We expect to see all logs from namespace `foo` only be in tenant `foo` and all logs from namespace `bar` only be in tenant `bar`.
**What did you see instead?**
Logs from every namespace appear in other tenants.
below are some screenshots from Grafana where we expect to only see logs from the `foo` namespace.

Instead, we see many logs from other namespaces as well (even though our X-Scope-OrgID is set to the `foo` tenant)

**What version did you use?**
Version: v0.57.2
**Additional context**
We noticed in testing environments that with small non-spiky loads the logs were grouped and exported to the correct tenant, but in high-volume spiky environments the logs were not.
|
process
|
resource attributes sent to incorrect tenant describe the bug multi tenant use cases in the loki exporter were introduced through which allows setting and creating tenantids from resource attribute values this is awesome and is the exact functionality we are looking for but we are somehow seeing attribute values being sent to the wrong tenant especially when logging at scale for example if we configure the loki exporter to set the tenantid from the namespace attribute we are seeing many logs from namespace foo show up in the namespace bar tenant in loki steps to reproduce this has been occurring using the opentelemetry collector helm chart with the following configuration we are following the recommendations in the documentation to use the groupbyattrs processor as well as batching yaml config receivers fluentforward endpoint processors groupbyattrs keys namespace name filter exclude logs exclude match type regexp resource attributes key namespace name value istio system kube system exporters loki endpoint format json tenant source attributes value namespace name labels resource namespace name attributes stream app pod name account cluster name region container name service pipelines logs exporters processors receivers to reproduce we noticed this occur more frequently during spiky loads at a higher volume but also noticed it occur sometimes when the opentelemetry collector was killed or rescheduled onto a new node as part of troubleshooting we have also tested multiple combinations of this config such as batching and grouping after the filter as well as before and also not batching at all none of these have resulted in proper tenants what did you expect to see we expect to see all logs from namespace foo only be in tenant foo and all logs from namespace bar only be in tenant bar what did you see instead logs from every namespace appear in other tenants below are some screenshots from grafana where we expect to only see logs from the foo namespace instead we see many logs from other namespaces as well even though our x scope orgid is set to the foo tenant what version did you use version additional context we noticed in testing environments that with small non spiky loads the logs were grouped and exported to the correct tenant but in high volume spiky environments the logs were not
| 1
|
446,263
| 31,465,820,958
|
IssuesEvent
|
2023-08-30 01:50:59
|
rosrob/TecnoXpress
|
https://api.github.com/repos/rosrob/TecnoXpress
|
opened
|
IEEE 830 3 Requisitos específicos
|
documentation
|
Completar el punto 3 de requisitos especificos
_Product Backlog
Escribir todas las historias de Usuarios encontradas para el proyecto_
|
1.0
|
IEEE 830 3 Requisitos específicos - Completar el punto 3 de requisitos especificos
_Product Backlog
Escribir todas las historias de Usuarios encontradas para el proyecto_
|
non_process
|
ieee requisitos específicos completar el punto de requisitos especificos product backlog escribir todas las historias de usuarios encontradas para el proyecto
| 0
|
5,655
| 8,526,549,238
|
IssuesEvent
|
2018-11-02 16:35:12
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Handle leak while creating processes simultaneously from two threads causes spurious deadlocks
|
area-System.Diagnostics.Process bug
|
The fundamental issue is the same on all platforms. The varying behaviors are due to technical details of the platforms.
On Windows, the following leaks are possible:
When creating two processes that redirect handles, standard IO reader and writer handles can leak from one process being created to another process being created.
The following deadlock is possible:
`.ReadToEnd()` won't terminate because the standard output/error writer handle was inherited into another process that is waiting for input.
On Mac OSX the following leaks are possible:
When creating two processes that redirect handles, standard IO reader and writer handles can leak from one process being created to another process being created.
The following deadlocks are possible:
`.ReadToEnd()` won't terminate because the standard input writer was inherited into another process that is waiting for input.
`.ReadToEnd()` won't terminate because the standard output/error writer handle was inherited into another process that is waiting for input.
No leaks are possible on Linux because `pipe2()` is used correctly.
Solution, MAC OSX:
In https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs
// Lock to avoid races with OnSigChild
// By using a ReaderWriterLock we allow multiple processes to start concurrently.
#if HAVE_PIPE2
s_processStartLock.EnterReadLock();
#else
// Simultaneous create process doesn't work on Mac OSX because the lack of a pipe2() system call causes handles to leak from one process being created to another.
s_processStartLock.EnterWriteLock();
#endif
#if HAVE_PIPE2
s_processStartLock.ExitReadLock();
#else
s_processStartLock.ExitWriteLock();
#endif
On Windows, we can either take the same choice of putting an exclusive lock around the whole thing, or we can use STARTUPINFOEX as documented here: https://blogs.msdn.microsoft.com/oldnewthing/20111216-00/?p=8873
This is the last issue driving #306 ; when this is fixed #306 is made irrelevant and can be closed (kestrel would still break if using libuv and started simultaneously with create process in another thread but libuv is no longer the default).
|
1.0
|
Handle leak while creating processes simultaneously from two threads causes spurious deadlocks - The fundamental issue is the same on all platforms. The varying behaviors are due to technical details of the platforms.
On Windows, the following leaks are possible:
When creating two processes that redirect handles, standard IO reader and writer handles can leak from one process being created to another process being created.
The following deadlock is possible:
`.ReadToEnd()` won't terminate because the standard output/error writer handle was inherited into another process that is waiting for input.
On Mac OSX the following leaks are possible:
When creating two processes that redirect handles, standard IO reader and writer handles can leak from one process being created to another process being created.
The following deadlocks are possible:
`.ReadToEnd()` won't terminate because the standard input writer was inherited into another process that is waiting for input.
`.ReadToEnd()` won't terminate because the standard output/error writer handle was inherited into another process that is waiting for input.
No leaks are possible on Linux because `pipe2()` is used correctly.
Solution, MAC OSX:
In https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs
// Lock to avoid races with OnSigChild
// By using a ReaderWriterLock we allow multiple processes to start concurrently.
#if HAVE_PIPE2
s_processStartLock.EnterReadLock();
#else
// Simultaneous create process doesn't work on Mac OSX because the lack of a pipe2() system call causes handles to leak from one process being created to another.
s_processStartLock.EnterWriteLock();
#endif
#if HAVE_PIPE2
s_processStartLock.ExitReadLock();
#else
s_processStartLock.ExitWriteLock();
#endif
On Windows, we can either take the same choice of putting an exclusive lock around the whole thing, or we can use STARTUPINFOEX as documented here: https://blogs.msdn.microsoft.com/oldnewthing/20111216-00/?p=8873
This is the last issue driving #306 ; when this is fixed #306 is made irrelevant and can be closed (kestrel would still break if using libuv and started simultaneously with create process in another thread but libuv is no longer the default).
|
process
|
handle leak while creating processes simultaneously from two threads causes spurious deadlocks the fundamental issue is the same on all platforms the varying behaviors are due to technical details of the platforms on windows the following leaks are possible when creating two processes that redirect handles standard io reader and writer handles can leak from one process being created to another process being created the following deadlock is possible readtoend won t terminate because the standard output error writer handle was inherited into another process that is waiting for input on mac osx the following leaks are possible when creating two processes that redirect handles standard io reader and writer handles can leak from one process being created to another process being created the following deadlocks are possible readtoend won t terminate because the standard input writer was inherited into another process that is waiting for input readtoend won t terminate because the standard output error writer handle was inherited into another process that is waiting for input no leaks are possible on linux because is used correctly solution mac osx in lock to avoid races with onsigchild by using a readerwriterlock we allow multiple processes to start concurrently if have s processstartlock enterreadlock else simultaneous create process doesn t work on mac osx because the lack of a system call causes handles to leak from one process being created to another s processstartlock enterwritelock endif if have s processstartlock exitreadlock else s processstartlock exitwritelock endif on windows we can either take the same choice of putting an exclusive lock around the whole thing or we can use startupinfoex as documented here this is the last issue driving when this is fixed is made irrelevant and can be closed kestrel would still break if using libuv and started simultaneously with create process in another thread but libuv is no longer the default
| 1
|
5,217
| 8,016,437,940
|
IssuesEvent
|
2018-07-25 13:22:12
|
w3c/transitions
|
https://api.github.com/repos/w3c/transitions
|
opened
|
How is "description" used in publications?
|
Process Issue
|
The publication process, for instance https://www.w3.org/Guide/transitions?profile=WG-NOTE#pubreq, requires a description to be included along with publication requests. Folks do this but aren't sure where this description is used. (cc @deniak )
|
1.0
|
How is "description" used in publications? - The publication process, for instance https://www.w3.org/Guide/transitions?profile=WG-NOTE#pubreq, requires a description to be included along with publication requests. Folks do this but aren't sure where this description is used. (cc @deniak )
|
process
|
how is description used in publications the publication process for instance requires a description to be included along with publication requests folks do this but aren t sure where this description is used cc deniak
| 1
|
35,261
| 9,554,802,270
|
IssuesEvent
|
2019-05-02 23:32:59
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
Report modifications needed for loan form
|
CF Report Builder Priority-High
|
I need the following loan form to be duplicated as a new form in Arctos reports, called "MSB Zoo Loan Invoice" with the following changes:
1) replace NK as the identifier with the ID "Albuquerque Biopark Zoo Local ID", and call the column "Local ID / RIO GRAND"
2) Add a column for the data of collection
<!-- probot = {"768919":{"who":"Jegelewicz","what":"","when":"2019-05-02T09:00:00.000Z"}} -->
|
1.0
|
Report modifications needed for loan form - I need the following loan form to be duplicated as a new form in Arctos reports, called "MSB Zoo Loan Invoice" with the following changes:
1) replace NK as the identifier with the ID "Albuquerque Biopark Zoo Local ID", and call the column "Local ID / RIO GRAND"
2) Add a column for the data of collection
<!-- probot = {"768919":{"who":"Jegelewicz","what":"","when":"2019-05-02T09:00:00.000Z"}} -->
|
non_process
|
report modifications needed for loan form i need the following loan form to be duplicated as a new form in arctos reports called msb zoo loan invoice with the following changes replace nk as the identifier with the id albuquerque biopark zoo local id and call the column local id rio grand add a column for the data of collection
| 0
|
6,422
| 9,526,712,367
|
IssuesEvent
|
2019-04-28 22:17:30
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Unbundle mobile tools
|
P1 team-Android type: process
|
We're about to add another tool for Android builds to do sharded dex compilation, and that depends on having an Android SDK available. If we put it into @bazel_tools, then we'd implicitly require every user of Bazel to also have an Android SDK, whether they work with Android or not.
After talking to @ahumesky and others, it seems to me that the right solution is to unbundle the Android tools - all of them. This requires that all devs who work with Android now also need to depend on the android tools repository (@android_tools?), but that's better than the alternative.
I think we have to do this before we can declare a 1.0, and if we want to ship further improvements to Android builds, sooner rather than later.
|
1.0
|
Unbundle mobile tools - We're about to add another tool for Android builds to do sharded dex compilation, and that depends on having an Android SDK available. If we put it into @bazel_tools, then we'd implicitly require every user of Bazel to also have an Android SDK, whether they work with Android or not.
After talking to @ahumesky and others, it seems to me that the right solution is to unbundle the Android tools - all of them. This requires that all devs who work with Android now also need to depend on the android tools repository (@android_tools?), but that's better than the alternative.
I think we have to do this before we can declare a 1.0, and if we want to ship further improvements to Android builds, sooner rather than later.
|
process
|
unbundle mobile tools we re about to add another tool for android builds to do sharded dex compilation and that depends on having an android sdk available if we put it into bazel tools then we d implicitly require every user of bazel to also have an android sdk whether they work with android or not after talking to ahumesky and others it seems to me that the right solution is to unbundle the android tools all of them this requires that all devs who work with android now also need to depend on the android tools repository android tools but that s better than the alternative i think we have to do this before we can declare a and if we want to ship further improvements to android builds sooner rather than later
| 1
|
6,429
| 9,531,331,325
|
IssuesEvent
|
2019-04-29 15:45:51
|
googleapis/google-cloud-java
|
https://api.github.com/repos/googleapis/google-cloud-java
|
closed
|
Update new_client.py to add new clients to the list of modules in google-cloud-clients
|
type: process
|
New clients should be added to the list of modules in the pom.xml file.
|
1.0
|
Update new_client.py to add new clients to the list of modules in google-cloud-clients - New clients should be added to the list of modules in the pom.xml file.
|
process
|
update new client py to add new clients to the list of modules in google cloud clients new clients should be added to the list of modules in the pom xml file
| 1
|
17,934
| 23,932,611,117
|
IssuesEvent
|
2022-09-10 19:28:14
|
ShawnPotter/snek-bot
|
https://api.github.com/repos/ShawnPotter/snek-bot
|
opened
|
Refactor Code
|
Ongoing Process
|
Everything is currently on index.js. Need a directory system to start separating out different functionality.
|
1.0
|
Refactor Code - Everything is currently on index.js. Need a directory system to start separating out different functionality.
|
process
|
refactor code everything is currently on index js need a directory system to start separating out different functionality
| 1
|
17,194
| 22,772,944,373
|
IssuesEvent
|
2022-07-08 11:51:30
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Obsoletion notice: GO:0039508 suppression by virus of host receptor activity & children
|
obsoletion multi-species process
|
Dear all,
The proposal has been made to obsolete GO:0039508 suppression by virus of host receptor activity & children
* GO:0039511 suppression by virus of host interferon receptor activity
* GO:0039509 suppression by virus of host pattern recognition receptor activity
The reason for obsoletion is that these terms represent molecular functions. There are no annotations to these terms; these terms are not in any subsets. There is one UniProt Keyword mapping: UniProtKB-KW:KW-1091 Inhibition of host interferon receptors by virus > GO:0039511:suppression by virus of host interferon receptor activity ; this can be remapped to the new term GO:0140886 suppression by virus of host interferon-mediated signaling pathway.
You can comment on the ticket: https://github.com/geneontology/go-ontology/issues/23649
Thanks, Pascale
|
1.0
|
Obsoletion notice: GO:0039508 suppression by virus of host receptor activity & children - Dear all,
The proposal has been made to obsolete GO:0039508 suppression by virus of host receptor activity & children
* GO:0039511 suppression by virus of host interferon receptor activity
* GO:0039509 suppression by virus of host pattern recognition receptor activity
The reason for obsoletion is that these terms represent molecular functions. There are no annotations to these terms; these terms are not in any subsets. There is one UniProt Keyword mapping: UniProtKB-KW:KW-1091 Inhibition of host interferon receptors by virus > GO:0039511:suppression by virus of host interferon receptor activity ; this can be remapped to the new term GO:0140886 suppression by virus of host interferon-mediated signaling pathway.
You can comment on the ticket: https://github.com/geneontology/go-ontology/issues/23649
Thanks, Pascale
|
process
|
obsoletion notice go suppression by virus of host receptor activity children dear all the proposal has been made to obsolete go suppression by virus of host receptor activity children go suppression by virus of host interferon receptor activity go suppression by virus of host pattern recognition receptor activity the reason for obsoletion is that these terms represent molecular functions there are no annotations to these terms these terms are not in any subsets there is one uniprot keyword mapping uniprotkb kw kw inhibition of host interferon receptors by virus go suppression by virus of host interferon receptor activity this can be remapped to the new term go suppression by virus of host interferon mediated signaling pathway you can comment on the ticket thanks pascale
| 1
|
383,785
| 11,362,376,657
|
IssuesEvent
|
2020-01-26 20:52:54
|
TNelen/BlackBox
|
https://api.github.com/repos/TNelen/BlackBox
|
opened
|
UI improvements
|
PRIORITY UI
|
Vote knop en start knop meer naar boven + meer marge tussen rand en items
Knoppen ronder maken, icoontjes kleiner
Bij lobby tekst bovenaan vet zetten
|
1.0
|
UI improvements - Vote knop en start knop meer naar boven + meer marge tussen rand en items
Knoppen ronder maken, icoontjes kleiner
Bij lobby tekst bovenaan vet zetten
|
non_process
|
ui improvements vote knop en start knop meer naar boven meer marge tussen rand en items knoppen ronder maken icoontjes kleiner bij lobby tekst bovenaan vet zetten
| 0
|
114,511
| 24,610,336,076
|
IssuesEvent
|
2022-10-14 20:39:23
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
closed
|
--send-logs and --log-level options in clickhouse-disks tool do not work
|
unfinished code
|
**Description**
--send-logs and --log-level options are present in the help, but they do not work.
```
# clickhouse-disks --help
...
--send-logs Send logs
--log-level Logging level
```
```
# clickhouse-disks --log-level ERROR list-disks
Bad arguments: too many positional options have been specified on the command line
```
```
# clickhouse-disks --send-logs false list-disks
Bad arguments: too many positional options have been specified on the command line
```
ClickHouse version: 22.8.6.71
**Use case**
Looking for a way to disable logging to stdout.
|
1.0
|
--send-logs and --log-level options in clickhouse-disks tool do not work - **Description**
--send-logs and --log-level options are present in the help, but they do not work.
```
# clickhouse-disks --help
...
--send-logs Send logs
--log-level Logging level
```
```
# clickhouse-disks --log-level ERROR list-disks
Bad arguments: too many positional options have been specified on the command line
```
```
# clickhouse-disks --send-logs false list-disks
Bad arguments: too many positional options have been specified on the command line
```
ClickHouse version: 22.8.6.71
**Use case**
Looking for a way to disable logging to stdout.
|
non_process
|
send logs and log level options in clickhouse disks tool do not work description send logs and log level options are present in the help but they do not work clickhouse disks help send logs send logs log level logging level clickhouse disks log level error list disks bad arguments too many positional options have been specified on the command line clickhouse disks send logs false list disks bad arguments too many positional options have been specified on the command line clickhouse version use case looking for a way to disable logging to stdout
| 0
|
244,547
| 20,676,582,771
|
IssuesEvent
|
2022-03-10 09:51:56
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
reopened
|
Longhorn App fails to install from Legacy project app
|
[zube]: To Test area/apps status/release-blocker kind/bug-qa team/area3
|
<!--------- For bugs and general issues --------->
**Setup**
- Rancher version: v2.6-head - `3f30288`
- Browser type & version: Chrome 97.0.4692.99
**Describe the bug**
Longhorn app can't be installed from legacy project app
**To Reproduce**
1. Enable legacy flag
2. Go to Legacy --> Project - > app
3. Launch Longhorn version v1.2.3
**Result**
Longhorn fails to install. No events, no pods get created.
**Expected Result**
Longhorn should be installed properly.
**Screenshots**
<img width="919" alt="Screen Shot 2022-02-16 at 11 40 39 AM" src="https://user-images.githubusercontent.com/60111667/154353548-f6a43ace-dfdb-4015-8f5a-9c9d1b72b8a4.png">
|
1.0
|
Longhorn App fails to install from Legacy project app - <!--------- For bugs and general issues --------->
**Setup**
- Rancher version: v2.6-head - `3f30288`
- Browser type & version: Chrome 97.0.4692.99
**Describe the bug**
Longhorn app can't be installed from legacy project app
**To Reproduce**
1. Enable legacy flag
2. Go to Legacy --> Project - > app
3. Launch Longhorn version v1.2.3
**Result**
Longhorn fails to install. No events, no pods get created.
**Expected Result**
Longhorn should be installed properly.
**Screenshots**
<img width="919" alt="Screen Shot 2022-02-16 at 11 40 39 AM" src="https://user-images.githubusercontent.com/60111667/154353548-f6a43ace-dfdb-4015-8f5a-9c9d1b72b8a4.png">
|
non_process
|
longhorn app fails to install from legacy project app setup rancher version head browser type version chrome describe the bug longhorn app can t be installed from legacy project app to reproduce enable legacy flag go to legacy project app launch longhorn version result longhorn fails to install no events no pods get created expected result longhorn should be installed properly screenshots img width alt screen shot at am src
| 0
|
4,077
| 7,017,456,585
|
IssuesEvent
|
2017-12-21 09:43:11
|
openvstorage/framework-alba-plugin
|
https://api.github.com/repos/openvstorage/framework-alba-plugin
|
closed
|
The role check when adding a backend does not refresh
|
priority_minor process_duplicate type_bug
|
When having the 'Add Backend' wizard open without having a DB role assigned to any StorageRouter, the role check in the wizard will fail and not allow to add a Backend. When adding a DB role via the API or any other way without closing the 'Add Backend' wizard, the check does not refresh
|
1.0
|
The role check when adding a backend does not refresh - When having the 'Add Backend' wizard open without having a DB role assigned to any StorageRouter, the role check in the wizard will fail and not allow to add a Backend. When adding a DB role via the API or any other way without closing the 'Add Backend' wizard, the check does not refresh
|
process
|
the role check when adding a backend does not refresh when having the add backend wizard open without having a db role assigned to any storagerouter the role check in the wizard will fail and not allow to add a backend when adding a db role via the api or any other way without closing the add backend wizard the check does not refresh
| 1
|
51,782
| 6,548,453,330
|
IssuesEvent
|
2017-09-04 22:04:41
|
InfiniteFlightAirportEditing/Airports
|
https://api.github.com/repos/InfiniteFlightAirportEditing/Airports
|
closed
|
LSGG-Geneve Airport-GENEVE-SWITZERLAND
|
Being Redone Design Issue Global Priority 2 (+10k)
|
Doing this one in the future when done with region airports.
|
1.0
|
LSGG-Geneve Airport-GENEVE-SWITZERLAND - Doing this one in the future when done with region airports.
|
non_process
|
lsgg geneve airport geneve switzerland doing this one in the future when done with region airports
| 0
|
15,740
| 19,910,494,140
|
IssuesEvent
|
2022-01-25 16:41:48
|
input-output-hk/high-assurance-legacy
|
https://api.github.com/repos/input-output-hk/high-assurance-legacy
|
closed
|
Formally prove `multi_receive` core lemmas
|
type: enhancement language: isabelle topic: process calculus
|
Our goal is to formally prove the `multi_receive` core lemmas described in #32.
Informal proofs of these lemmas are given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
|
1.0
|
Formally prove `multi_receive` core lemmas - Our goal is to formally prove the `multi_receive` core lemmas described in #32.
Informal proofs of these lemmas are given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
|
process
|
formally prove multi receive core lemmas our goal is to formally prove the multi receive core lemmas described in informal proofs of these lemmas are given in
| 1
|
172,641
| 6,514,925,198
|
IssuesEvent
|
2017-08-26 08:06:57
|
Bartlebys/Bartleby
|
https://api.github.com/repos/Bartlebys/Bartleby
|
closed
|
Interactive Quarantines Clean up Procedure
|
feature High Priority
|
When an operation is in quarantine we should present a Synthesis to the user and Ask what to do.
- A **synthesis** will explain the problem with the operation (Deletion of part of A Bunch Update, ACL issue).
- The system will offer solutions candidats like : "Delete entity X locally."
- The user must arbitrate or defer the Arbitration
To understand what is Operation Quarantine [check the documentation](https://github.com/Bartlebys/Bartleby/blob/master/DataSynchronization.md)
|
1.0
|
Interactive Quarantines Clean up Procedure - When an operation is in quarantine we should present a Synthesis to the user and Ask what to do.
- A **synthesis** will explain the problem with the operation (Deletion of part of A Bunch Update, ACL issue).
- The system will offer solutions candidats like : "Delete entity X locally."
- The user must arbitrate or defer the Arbitration
To understand what is Operation Quarantine [check the documentation](https://github.com/Bartlebys/Bartleby/blob/master/DataSynchronization.md)
|
non_process
|
interactive quarantines clean up procedure when an operation is in quarantine we should present a synthesis to the user and ask what to do a synthesis will explain the problem with the operation deletion of part of a bunch update acl issue the system will offer solutions candidats like delete entity x locally the user must arbitrate or defer the arbitration to understand what is operation quarantine
| 0
|
177,660
| 13,735,052,073
|
IssuesEvent
|
2020-10-05 09:34:55
|
topcoder-platform/community-app
|
https://api.github.com/repos/topcoder-platform/community-app
|
closed
|
Challenge details page displaying a blank screen
|
P0 QA Pass on Test Staging Env listing:re-implementation
|
Description:
Challenge details page displaying a blank screen

|
1.0
|
Challenge details page displaying a blank screen - Description:
Challenge details page displaying a blank screen

|
non_process
|
challenge details page displaying a blank screen description challenge details page displaying a blank screen
| 0
|
19,365
| 25,494,303,950
|
IssuesEvent
|
2022-11-27 13:34:49
|
googleapis/python-documentai
|
https://api.github.com/repos/googleapis/python-documentai
|
closed
|
Internal error encountered with sample test_batch_process_documents_with_bad_input
|
status: investigating type: process priority: p2 external api: documentai samples
|
In the last 2 days, the samples test [test_batch_process_documents_with_bad_input](https://github.com/googleapis/python-documentai/blob/main/samples/snippets/batch_process_documents_sample_bad_input_test.py#L30) has been failing with
```
> assert "Failed to process" in e.message
E AssertionError: assert 'Failed to process' in 'Internal error encountered.'
E + where 'Internal error encountered.' = InternalServerError('Internal error encountered.',).message
```
See the build log [here](https://source.cloud.google.com/results/invocations/aba8b613-3e3b-4d55-87f5-598842a7b8a3/log).
|
1.0
|
Internal error encountered with sample test_batch_process_documents_with_bad_input - In the last 2 days, the samples test [test_batch_process_documents_with_bad_input](https://github.com/googleapis/python-documentai/blob/main/samples/snippets/batch_process_documents_sample_bad_input_test.py#L30) has been failing with
```
> assert "Failed to process" in e.message
E AssertionError: assert 'Failed to process' in 'Internal error encountered.'
E + where 'Internal error encountered.' = InternalServerError('Internal error encountered.',).message
```
See the build log [here](https://source.cloud.google.com/results/invocations/aba8b613-3e3b-4d55-87f5-598842a7b8a3/log).
|
process
|
internal error encountered with sample test batch process documents with bad input in the last days the samples test has been failing with assert failed to process in e message e assertionerror assert failed to process in internal error encountered e where internal error encountered internalservererror internal error encountered message see the build log
| 1
|
5,230
| 8,030,162,756
|
IssuesEvent
|
2018-07-27 18:35:01
|
GerkeLab/curatedProstateData
|
https://api.github.com/repos/GerkeLab/curatedProstateData
|
opened
|
Data types for collection
|
Process :clapper:
|
From Dr. Costello, July 19, 2018
> We should agree upon...the data types we are collecting. Gene expression will likely be the data type most likely to span most data sets. I think collecting any other -omic data type would be useful as well.
|
1.0
|
Data types for collection - From Dr. Costello, July 19, 2018
> We should agree upon...the data types we are collecting. Gene expression will likely be the data type most likely to span most data sets. I think collecting any other -omic data type would be useful as well.
|
process
|
data types for collection from dr costello july we should agree upon the data types we are collecting gene expression will likely be the data type most likely to span most data sets i think collecting any other omic data type would be useful as well
| 1
|
799,507
| 28,308,390,411
|
IssuesEvent
|
2023-04-10 13:17:16
|
sekiguchi-nagisa/ydsh
|
https://api.github.com/repos/sekiguchi-nagisa/ydsh
|
closed
|
show suggestion for semantic error
|
enhancement Core priority:high
|
show the following suggestion
* variable name
* type name
* method/field name
|
1.0
|
show suggestion for semantic error - show the following suggestion
* variable name
* type name
* method/field name
|
non_process
|
show suggestion for semantic error show the following suggestion variable name type name method field name
| 0
|
832,203
| 32,076,046,623
|
IssuesEvent
|
2023-09-25 11:04:49
|
GSM-MSG/Hi-v2-BackEnd
|
https://api.github.com/repos/GSM-MSG/Hi-v2-BackEnd
|
opened
|
QueryDsl 세팅
|
1️⃣ Priority: High ♻️ Refactor
|
### Describe
학생 전체 조회를 학번순으로 반환해주기 위해 queryDsl 을 사용하겠습니다.
### Additional
_No response_
|
1.0
|
QueryDsl 세팅 - ### Describe
학생 전체 조회를 학번순으로 반환해주기 위해 queryDsl 을 사용하겠습니다.
### Additional
_No response_
|
non_process
|
querydsl 세팅 describe 학생 전체 조회를 학번순으로 반환해주기 위해 querydsl 을 사용하겠습니다 additional no response
| 0
|
1,446
| 4,020,056,153
|
IssuesEvent
|
2016-05-16 17:02:31
|
emergence-lab/emergence-lab
|
https://api.github.com/repos/emergence-lab/emergence-lab
|
closed
|
Process Title/Description
|
backend enhancement process
|
Rework process model to have title and description to encourage better practices in users, similar to how git commits work
|
1.0
|
Process Title/Description - Rework process model to have title and description to encourage better practices in users, similar to how git commits work
|
process
|
process title description rework process model to have title and description to encourage better practices in users similar to how git commits work
| 1
|
164,827
| 26,029,025,163
|
IssuesEvent
|
2022-12-21 19:07:50
|
kubeshop/tracetest
|
https://api.github.com/repos/kubeshop/tracetest
|
closed
|
[In-App Config] Setup Wizard
|
design frontend backend
|
The setup wizard is the main in-app config feature that allows users to configure tracetest to start supporting traces.
Acceptance Criteria:
AC1:
As a user looking at the setup wizard
I should be able to see a brief explanation about why a data store is needed
And a way to select the data store of my preference
AC2 (Special Case for Otel Collector):
As a user looking at the UI setup wizard
And I select the Otel Collector as my data store
I should see a small tutorial with the steps I need to take to configure it
AC3 (Data Store Step):
https://github.com/kubeshop/tracetest/issues/1627
AC7:
As a user looking at the UI setup wizard
After filling in the details to configure the data store
I should be able to save the changes
Without testing the connection
AC8:
As a user looking at the UI setup wizard
And I have clicked save changes
I should see a prompt message letting me know that Tracetest will be restarted and there will be some downtime after confirmation.
AC9:
As a user that just went through the setup process
I should see a change in the UI that will tell me that Tracetest is ready to support traces
|
1.0
|
[In-App Config] Setup Wizard - The setup wizard is the main in-app config feature that allows users to configure tracetest to start supporting traces.
Acceptance Criteria:
AC1:
As a user looking at the setup wizard
I should be able to see a brief explanation about why a data store is needed
And a way to select the data store of my preference
AC2 (Special Case for Otel Collector):
As a user looking at the UI setup wizard
And I select the Otel Collector as my data store
I should see a small tutorial with the steps I need to take to configure it
AC3 (Data Store Step):
https://github.com/kubeshop/tracetest/issues/1627
AC7:
As a user looking at the UI setup wizard
After filling in the details to configure the data store
I should be able to save the changes
Without testing the connection
AC8:
As a user looking at the UI setup wizard
And I have clicked save changes
I should see a prompt message letting me know that Tracetest will be restarted and there will be some downtime after confirmation.
AC9:
As a user that just went through the setup process
I should see a change in the UI that will tell me that Tracetest is ready to support traces
|
non_process
|
setup wizard the setup wizard is the main in app config feature that allows users to configure tracetest to start supporting traces acceptance criteria as a user looking at the setup wizard i should be able to see a brief explanation about why a data store is needed and a way to select the data store of my preference special case for otel collector as a user looking at the ui setup wizard and i select the otel collector as my data store i should see a small tutorial with the steps i need to take to configure it data store step as a user looking at the ui setup wizard after filling in the details to configure the data store i should be able to save the changes without testing the connection as a user looking at the ui setup wizard and i have clicked save changes i should see a prompt message letting me know that tracetest will be restarted and there will be some downtime after confirmation as a user that just went through the setup process i should see a change in the ui that will tell me that tracetest is ready to support traces
| 0
|
39,748
| 16,088,293,754
|
IssuesEvent
|
2021-04-26 13:54:44
|
Ryujinx/Ryujinx
|
https://api.github.com/repos/Ryujinx/Ryujinx
|
opened
|
hid IIrSensorServer: 307 (RunClusteringProcessor) is not implemented
|
not-implemented service:hid
|
hid IIrSensorServer: 307 (RunClusteringProcessor) is not implemented
[Switchbrew](https://switchbrew.org/wiki/HID_services#RunClusteringProcessor) | [SwIPC](https://reswitched.github.io/SwIPC/ifaces.html#nn::irsensor::IIrSensorServer(307))
## Service description:
Takes a PID-descriptor, an #IrCameraHandle, an AppletResourceUserId, and a #PackedClusteringProcessorConfig. No output.
## Required by:
https://github.com/Ryujinx/Ryujinx-Games-List/issues/470 version 1.6.0 and above
|
1.0
|
hid IIrSensorServer: 307 (RunClusteringProcessor) is not implemented - hid IIrSensorServer: 307 (RunClusteringProcessor) is not implemented
[Switchbrew](https://switchbrew.org/wiki/HID_services#RunClusteringProcessor) | [SwIPC](https://reswitched.github.io/SwIPC/ifaces.html#nn::irsensor::IIrSensorServer(307))
## Service description:
Takes a PID-descriptor, an #IrCameraHandle, an AppletResourceUserId, and a #PackedClusteringProcessorConfig. No output.
## Required by:
https://github.com/Ryujinx/Ryujinx-Games-List/issues/470 version 1.6.0 and above
|
non_process
|
hid iirsensorserver runclusteringprocessor is not implemented hid iirsensorserver runclusteringprocessor is not implemented service description takes a pid descriptor an ircamerahandle an appletresourceuserid and a packedclusteringprocessorconfig no output required by version and above
| 0
|
19,567
| 25,887,827,304
|
IssuesEvent
|
2022-12-14 15:43:30
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_terminate_exit (__main__.SpawnTest)
|
high priority triage review module: multiprocessing triaged module: flaky-tests skipped
|
Platforms: linux
This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_terminate_exit%2C%20SpawnTest) and the most recent
[workflow logs](https://github.com/pytorch/pytorch/actions/runs/1817341053).
Over the past 6 hours, it has been determined flaky in 1 workflow(s) with
1 red and 3 green.
cc @ezyang @gchanan @zou3519 @VitalyFedyunin
|
1.0
|
DISABLED test_terminate_exit (__main__.SpawnTest) - Platforms: linux
This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_terminate_exit%2C%20SpawnTest) and the most recent
[workflow logs](https://github.com/pytorch/pytorch/actions/runs/1817341053).
Over the past 6 hours, it has been determined flaky in 1 workflow(s) with
1 red and 3 green.
cc @ezyang @gchanan @zou3519 @VitalyFedyunin
|
process
|
disabled test terminate exit main spawntest platforms linux this test was disabled because it is failing on trunk see and the most recent over the past hours it has been determined flaky in workflow s with red and green cc ezyang gchanan vitalyfedyunin
| 1
|
122,588
| 10,227,473,974
|
IssuesEvent
|
2019-08-16 20:54:11
|
mozilla/iris_firefox
|
https://api.github.com/repos/mozilla/iris_firefox
|
closed
|
Fix set_as_default_search_engine_using_one_off_context_menu test
|
test case
|
Can not reproduce failure described at step: "None - [Actual]: None [Expected]: None" for OSX
|
1.0
|
Fix set_as_default_search_engine_using_one_off_context_menu test - Can not reproduce failure described at step: "None - [Actual]: None [Expected]: None" for OSX
|
non_process
|
fix set as default search engine using one off context menu test can not reproduce failure described at step none none none for osx
| 0
|
56,215
| 6,506,931,912
|
IssuesEvent
|
2017-08-24 11:05:29
|
LDMW/cms
|
https://api.github.com/repos/LDMW/cms
|
closed
|
Feedback loop
|
please-test T1d
|
As a service provider, I want the service to ask returning visitors how they found the resources that they were directed to, so that I can start to gather valuable feedback on the resources on the service and continue to improve
- [x] If user has clicked on a resource, and then returns to LM, a section appears above the 'personalise your results' section which asks for feedback on the resource they went to
- [x] If they have clicked on multiple resources, multiple requests for feedback appear
- [x] Section appears across all pages on the site
- [x] Section doesn't disappear until the user has given feedback or closed the alert
- [x] Non JS alternative, the feedback loop won't appear for these users
- [x] Styled according to India's designs (COMING SOON)
|
1.0
|
Feedback loop - As a service provider, I want the service to ask returning visitors how they found the resources that they were directed to, so that I can start to gather valuable feedback on the resources on the service and continue to improve
- [x] If user has clicked on a resource, and then returns to LM, a section appears above the 'personalise your results' section which asks for feedback on the resource they went to
- [x] If they have clicked on multiple resources, multiple requests for feedback appear
- [x] Section appears across all pages on the site
- [x] Section doesn't disappear until the user has given feedback or closed the alert
- [x] Non JS alternative, the feedback loop won't appear for these users
- [x] Styled according to India's designs (COMING SOON)
|
non_process
|
feedback loop as a service provider i want the service to ask returning visitors how they found the resources that they were directed to so that i can start to gather valuable feedback on the resources on the service and continue to improve if user has clicked on a resource and then returns to lm a section appears above the personalise your results section which asks for feedback on the resource they went to if they have clicked on multiple resources multiple requests for feedback appear section appears across all pages on the site section doesn t disappear until the user has given feedback or closed the alert non js alternative the feedback loop won t appear for these users styled according to india s designs coming soon
| 0
|
266,659
| 20,160,648,979
|
IssuesEvent
|
2022-02-09 21:08:05
|
amosproj/amos2021ws03-teams-to-nextcloud
|
https://api.github.com/repos/amosproj/amos2021ws03-teams-to-nextcloud
|
closed
|
Github page documentation
|
documentation Est. Size = 3
|
## User story
1. As an administrator
2. I want to have a guide / manual which provides all important information about the app
3. So that I can understand how the app works
## Acceptance criteria
* Github page is created
* Guide provides information about the initial setup: installing the zip file, setting it up on teams and information about the configuration of the server (reference https://github.com/amosproj/amos2021ws03-teams-to-nextcloud/tree/main/Documentation)
## Definition of done
- [ ] Code is reviewed and merged to main by RM (or at least a different person from the one, who has implemented it)
- [ ] The code builds without errors or if there are any - they are noted down and clearly communicated with the team
- [ ] A feature is to be implemented so that it is scalable on different devices and works consistently for the rest of the team
- [ ] The code goes successfully through the CI pipeline
|
1.0
|
Github page documentation - ## User story
1. As an administrator
2. I want to have a guide / manual which provides all important information about the app
3. So that I can understand how the app works
## Acceptance criteria
* Github page is created
* Guide provides information about the initial setup: installing the zip file, setting it up on teams and information about the configuration of the server (reference https://github.com/amosproj/amos2021ws03-teams-to-nextcloud/tree/main/Documentation)
## Definition of done
- [ ] Code is reviewed and merged to main by RM (or at least a different person from the one, who has implemented it)
- [ ] The code builds without errors or if there are any - they are noted down and clearly communicated with the team
- [ ] A feature is to be implemented so that it is scalable on different devices and works consistently for the rest of the team
- [ ] The code goes successfully through the CI pipeline
|
non_process
|
github page documentation user story as an administrator i want to have a guide manual which provides all important information about the app so that i can understand how the app works acceptance criteria github page is created guide provides information about the initial setup installing the zip file setting it up on teams and information about the configuration of the server reference definition of done code is reviewed and merged to main by rm or at least a different person from the one who has implemented it the code builds without errors or if there are any they are noted down and clearly communicated with the team a feature is to be implemented so that it is scalable on different devices and works consistently for the rest of the team the code goes successfully through the ci pipeline
| 0
|
22,416
| 10,756,657,233
|
IssuesEvent
|
2019-10-31 11:41:40
|
perezLamed/lamed_flowchart
|
https://api.github.com/repos/perezLamed/lamed_flowchart
|
closed
|
WS-2019-0019 Medium Severity Vulnerability detected by WhiteSource
|
security vulnerability
|
## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>braces-0.1.5.tgz</b>, <b>braces-1.8.5.tgz</b></p></summary>
<p>
<details><summary><b>braces-0.1.5.tgz</b></p></summary>
<p>Fastest brace expansion lib. Typically used with file paths, but can be used with any string. Expands comma-separated values (e.g. `foo/{a,b,c}/bar`) and alphabetical or numerical ranges (e.g. `{1..9}`)</p>
<p>path: /tmp/WhiteSource-ArchiveExtractor_20190226220544_a1f5ea1f-7ab8-4472-bde9-a2375205cf16/20190226220544_5857/git_depth_0/lamed_Flowchart/dist/clipboard.js-master/clipboard.js-master/node_modules/expand-braces/node_modules/braces/package.json</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/braces/-/braces-0.1.5.tgz>http://registry.npmjs.org/braces/-/braces-0.1.5.tgz</a></p>
Dependency Hierarchy:
- karma-3.1.1.tgz (Root Library)
- expand-braces-0.1.2.tgz
- :x: **braces-0.1.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>path: /tmp/git/lamed_Flowchart/node_modules/lamed_io/node_modules/braces/package.json</p>
<p>
<p>Library home page: <a href=https://registry.npmjs.org/braces/-/braces-1.8.5.tgz>https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
Dependency Hierarchy:
- lamed_folder-0.1.0.tgz (Root Library)
- lamed_array-0.1.3.tgz
- expect-23.6.0.tgz
- jest-message-util-23.4.0.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-02-21
<p>URL: <a href=https://www.npmjs.com/advisories/786>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0019 Medium Severity Vulnerability detected by WhiteSource - ## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>braces-0.1.5.tgz</b>, <b>braces-1.8.5.tgz</b></p></summary>
<p>
<details><summary><b>braces-0.1.5.tgz</b></p></summary>
<p>Fastest brace expansion lib. Typically used with file paths, but can be used with any string. Expands comma-separated values (e.g. `foo/{a,b,c}/bar`) and alphabetical or numerical ranges (e.g. `{1..9}`)</p>
<p>path: /tmp/WhiteSource-ArchiveExtractor_20190226220544_a1f5ea1f-7ab8-4472-bde9-a2375205cf16/20190226220544_5857/git_depth_0/lamed_Flowchart/dist/clipboard.js-master/clipboard.js-master/node_modules/expand-braces/node_modules/braces/package.json</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/braces/-/braces-0.1.5.tgz>http://registry.npmjs.org/braces/-/braces-0.1.5.tgz</a></p>
Dependency Hierarchy:
- karma-3.1.1.tgz (Root Library)
- expand-braces-0.1.2.tgz
- :x: **braces-0.1.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>path: /tmp/git/lamed_Flowchart/node_modules/lamed_io/node_modules/braces/package.json</p>
<p>
<p>Library home page: <a href=https://registry.npmjs.org/braces/-/braces-1.8.5.tgz>https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
Dependency Hierarchy:
- lamed_folder-0.1.0.tgz (Root Library)
- lamed_array-0.1.3.tgz
- expect-23.6.0.tgz
- jest-message-util-23.4.0.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-02-21
<p>URL: <a href=https://www.npmjs.com/advisories/786>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium severity vulnerability detected by whitesource ws medium severity vulnerability vulnerable libraries braces tgz braces tgz braces tgz fastest brace expansion lib typically used with file paths but can be used with any string expands comma separated values e g foo a b c bar and alphabetical or numerical ranges e g path tmp whitesource archiveextractor git depth lamed flowchart dist clipboard js master clipboard js master node modules expand braces node modules braces package json library home page a href dependency hierarchy karma tgz root library expand braces tgz x braces tgz vulnerable library braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification path tmp git lamed flowchart node modules lamed io node modules braces package json library home page a href dependency hierarchy lamed folder tgz root library lamed array tgz expect tgz jest message util tgz micromatch tgz x braces tgz vulnerable library vulnerability details version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
3,740
| 6,733,147,628
|
IssuesEvent
|
2017-10-18 13:59:32
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Group Tag Action
|
enhancement form process workflow
|
Be able to Cancel and Un-cancel group of tags by given range and delimited list.
|
1.0
|
Group Tag Action - Be able to Cancel and Un-cancel group of tags by given range and delimited list.
|
process
|
group tag action be able to cancel and un cancel group of tags by given range and delimited list
| 1
|
630,636
| 20,115,271,059
|
IssuesEvent
|
2022-02-07 18:49:55
|
MGHComputationalPathology/slim
|
https://api.github.com/repos/MGHComputationalPathology/slim
|
opened
|
User-level error reporting is lacking
|
idc:priority
|
When I get a 404 error, there is nothing that gets communicated to the user in the interface - only in the JS console. I think this should be improved. What do you think @hackermd @Punzo ?

|
1.0
|
User-level error reporting is lacking - When I get a 404 error, there is nothing that gets communicated to the user in the interface - only in the JS console. I think this should be improved. What do you think @hackermd @Punzo ?

|
non_process
|
user level error reporting is lacking when i get a error there is nothing that gets communicated to the user in the interface only in the js console i think this should be improved what do you think hackermd punzo
| 0
|
22,117
| 30,646,215,234
|
IssuesEvent
|
2023-07-25 05:04:00
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
Prepare to release version v23.7.1
|
issue-processing-state-04
|
Update the version number in __init__.py for releasing the latest version of Quark.
In this version, the following changes will be included.
* #526
* #527
* #528
|
1.0
|
Prepare to release version v23.7.1 - Update the version number in __init__.py for releasing the latest version of Quark.
In this version, the following changes will be included.
* #526
* #527
* #528
|
process
|
prepare to release version update the version number in init py for releasing the latest version of quark in this version the following changes will be included
| 1
|
229,923
| 7,601,040,665
|
IssuesEvent
|
2018-04-28 09:09:35
|
metasfresh/metasfresh
|
https://api.github.com/repos/metasfresh/metasfresh
|
closed
|
Manufacturing Order automatic DocNo
|
priority:high type:bug
|
### Is this a bug or feature request?
Bug
### What is the current behavior?
In WebUI, the manufacturing order Document No is not created automatically.
#### Which are the steps to reproduce?
Open, try and see
### What is the expected or desired behavior?
Automatically create the manufacturing Order Document No.
|
1.0
|
Manufacturing Order automatic DocNo - ### Is this a bug or feature request?
Bug
### What is the current behavior?
In WebUI, the manufacturing order Document No is not created automatically.
#### Which are the steps to reproduce?
Open, try and see
### What is the expected or desired behavior?
Automatically create the manufacturing Order Document No.
|
non_process
|
manufacturing order automatic docno is this a bug or feature request bug what is the current behavior in webui the manufacturing order document no is not created automatically which are the steps to reproduce open try and see what is the expected or desired behavior automatically create the manufacturing order document no
| 0
|
34,184
| 4,892,569,174
|
IssuesEvent
|
2016-11-18 20:07:30
|
radare/radare2
|
https://api.github.com/repos/radare/radare2
|
opened
|
ej doesn't produce valid json
|
easy has-test json
|
ej doesn't produce valid json, several strings aren't escaped and hexadecimal aren't int:


|
1.0
|
ej doesn't produce valid json - ej doesn't produce valid json, several strings aren't escaped and hexadecimal aren't int:


|
non_process
|
ej doesn t produce valid json ej doesn t produce valid json several strings aren t escaped and hexadecimal aren t int
| 0
|
19
| 2,496,245,869
|
IssuesEvent
|
2015-01-06 18:06:31
|
vivo-isf/vivo-isf-ontology
|
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
|
closed
|
smooth muscle contraction
|
biological_process imported
|
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 17:29:23_
GO:0006939
parent: system process
\<a href="http://purl.obolibrary.org/obo/GO_0003008" rel="nofollow">http://purl.obolibrary.org/obo/GO_0003008</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=177_
|
1.0
|
smooth muscle contraction - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 17:29:23_
GO:0006939
parent: system process
\<a href="http://purl.obolibrary.org/obo/GO_0003008" rel="nofollow">http://purl.obolibrary.org/obo/GO_0003008</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=177_
|
process
|
smooth muscle contraction from on december go parent system process original issue
| 1
|
247,370
| 20,974,968,574
|
IssuesEvent
|
2022-03-28 14:31:12
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
[CI] DataTiersMigrationsTests testUserOptsOutOfTierMigration failing
|
>test-failure :Data Management/ILM+SLM Team:Data Management
|
**Build scan:**
https://gradle-enterprise.elastic.co/s/aodpvg6m2oba4/tests/:x-pack:plugin:ilm:internalClusterTest/org.elasticsearch.xpack.ilm.DataTiersMigrationsTests/testUserOptsOutOfTierMigration
**Reproduction line:**
`./gradlew ':x-pack:plugin:ilm:internalClusterTest' --tests "org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.testUserOptsOutOfTierMigration" -Dtests.seed=D3FF350541188AFA -Dtests.locale=ar-IQ -Dtests.timezone=Asia/Famagusta -Druntime.java=18`
**Applicable branches:**
8.1
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ilm.DataTiersMigrationsTests&tests.test=testUserOptsOutOfTierMigration
**Failure excerpt:**
```
java.lang.AssertionError:
Expected: is "warm"
but: was "new"
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.lambda$testUserOptsOutOfTierMigration$3(DataTiersMigrationsTests.java:192)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1070)
at org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.testUserOptsOutOfTierMigration(DataTiersMigrationsTests.java:187)
at jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
at java.lang.reflect.Method.invoke(Method.java:577)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:833)
```
|
1.0
|
[CI] DataTiersMigrationsTests testUserOptsOutOfTierMigration failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/aodpvg6m2oba4/tests/:x-pack:plugin:ilm:internalClusterTest/org.elasticsearch.xpack.ilm.DataTiersMigrationsTests/testUserOptsOutOfTierMigration
**Reproduction line:**
`./gradlew ':x-pack:plugin:ilm:internalClusterTest' --tests "org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.testUserOptsOutOfTierMigration" -Dtests.seed=D3FF350541188AFA -Dtests.locale=ar-IQ -Dtests.timezone=Asia/Famagusta -Druntime.java=18`
**Applicable branches:**
8.1
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ilm.DataTiersMigrationsTests&tests.test=testUserOptsOutOfTierMigration
**Failure excerpt:**
```
java.lang.AssertionError:
Expected: is "warm"
but: was "new"
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.lambda$testUserOptsOutOfTierMigration$3(DataTiersMigrationsTests.java:192)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1070)
at org.elasticsearch.xpack.ilm.DataTiersMigrationsTests.testUserOptsOutOfTierMigration(DataTiersMigrationsTests.java:187)
at jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
at java.lang.reflect.Method.invoke(Method.java:577)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:833)
```
|
non_process
|
datatiersmigrationstests testuseroptsoutoftiermigration failing build scan reproduction line gradlew x pack plugin ilm internalclustertest tests org elasticsearch xpack ilm datatiersmigrationstests testuseroptsoutoftiermigration dtests seed dtests locale ar iq dtests timezone asia famagusta druntime java applicable branches reproduces locally no failure history failure excerpt java lang assertionerror expected is warm but was new at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch xpack ilm datatiersmigrationstests lambda testuseroptsoutoftiermigration datatiersmigrationstests java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch xpack ilm datatiersmigrationstests testuseroptsoutoftiermigration datatiersmigrationstests java at jdk internal reflect directmethodhandleaccessor invoke directmethodhandleaccessor java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
| 0
|
20,201
| 26,778,158,190
|
IssuesEvent
|
2023-01-31 18:48:31
|
Ravenbrook/mps
|
https://api.github.com/repos/Ravenbrook/mps
|
closed
|
Merge procedure creates bogus Perforce branch locations
|
git-migration process optional
|
At some point we gained a bogus top-level directory in Perforce at //info.ravenbrook.com/project/mps/save-errno-win32. How'd that happen? It's something that's very hard to correct. How do we prevent it happening again?
Basically, this happened because a branch was pushed that did not conform to the branch/DATE/TOPIC naming convention. Git Fusion is mapping branch names to Perforce directories relative to //info.ravenbrook.com/project/mps, so if a branch is pushed called "foo" it'll end up at //info.ravenbrook.com/project/mps/foo/...
Here's how I found out what happened and my hypothesis of the steps that led to this.
The branch was created by Git Fusion and changes within it belong to the placeholder user "unknown_git":
```
$ p4 changes save-errno-win32/...
Change 197885 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Fixed typos in the test/functio'
Change 197884 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Save and restore GetLastError()'
Change 197883 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Added a test case for GitHub is'
Change 197882 on 2021/08/08 by git-fusion-user@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Git Fusion branch management I'
```
When Git Fusion creates a branch like this, it updates the configuration for the repo, so we can track it down there:
```
$ p4 annotate -c //.git-fusion/repos/mps-public/p4gf_config | grep save-errno-win32
197888: git-branch-name = save-errno-win32
...
$ p4 changes //...@197886,197888
Change 197888 on 2021/08/08 by git-fusion-user@git-fusion--perforce-8226a998-f83a-11eb-8966-00163e71ffc3 'Update 'p4gf_config'. '
Change 197887 on 2021/08/08 by git-fusion-user@git-fusion--perforce-8226a998-f83a-11eb-8966-00163e71ffc3 'Git Fusion 'mps-public' copied '
Change 197886 on 2021/08/08 by gdr@git-fusion--temp-perforce-mps-public-82b3ee86-f83a-11eb-b986-00163e71ffc3 'Merge branch 'save-errno-win32''
```
So we can see that this was caused by a merge of a third-party pull request by @gareth-rees :
```
$ p4 describe -s 197886
Change 197886 by gdr@git-fusion--temp-perforce-mps-public-82b3ee86-f83a-11eb-b986-00163e71ffc3 on 2021/08/08 12:19:50
Merge branch 'save-errno-win32' from Filip Strömbäck.
Imported from Git
Author: Gareth Rees [<gdr@garethrees.org>](mailto:gdr@garethrees.org) 1628421098 +0100 (2021-08-08 12:11:38+01:00)
Committer: Gareth Rees [<gdr@garethrees.org>](mailto:gdr@garethrees.org) 1628421098 +0100 (2021-08-08 12:11:38+01:00)
sha1: 86a528d1cdee81ecbf9365e3db7a65b1a3aaa05b
push-state: complete
parent-changes: cb4051bd02a61cc75927bb28aa4d2643a848c3eb=[197876, 197878]/8360988487490952f6edae36b70c1bba00fe6d2f=[197885]
Affected files ...
... //info.ravenbrook.com/project/mps/master/code/protw3.c#31 edit
... //info.ravenbrook.com/project/mps/master/design/thread-manager.txt#26 edit
... //info.ravenbrook.com/project/mps/master/test/function/237.c#1 add
```
This explains the "unknown_git" changes: they were made by @fstromback .
@fstromback's branch was imported into the git repo and can be see at https://github.com/Ravenbrook/mps/tree/save-errno-win32. How'd that happen?
There is no pull request by @fstromback by searching https://github.com/Ravenbrook/mps/pulls?q=is%3Apr+fstromback
This is where we can see flaws in GitHub's system. How do we track this down? Searching for a branch name in the pull requests does not reveal them. (I tested with some known ones.) I just had to make some guesses and found https://github.com/Ravenbrook/mps/pull/62
That pull request resulted in a merge of a branch from a fork of the MPS at https://github.com/fstromback/mps/tree/save-errno-win32 .
Somehow (probably GitHub's default behaviour via the GUI) this created a branch of the same name in Ravenbrook's repo at https://github.com/Ravenbrook/mps/tree/save-errno-win32.
From there I think @gareth-rees followed [his procedure](https://info.ravenbrook.com/mail/2020/09/03/13-02-35/0/) and at step 7 did `git push ravenbrook master "$BRANCH”` at which point Git Fusion created //info.ravenbrook.com/project/mps/save-errno-win32.
|
1.0
|
Merge procedure creates bogus Perforce branch locations - At some point we gained a bogus top-level directory in Perforce at //info.ravenbrook.com/project/mps/save-errno-win32. How'd that happen? It's something that's very hard to correct. How do we prevent it happening again?
Basically, this happened because a branch was pushed that did not conform to the branch/DATE/TOPIC naming convention. Git Fusion is mapping branch names to Perforce directories relative to //info.ravenbrook.com/project/mps, so if a branch is pushed called "foo" it'll end up at //info.ravenbrook.com/project/mps/foo/...
Here's how I found out what happened and my hypothesis of the steps that led to this.
The branch was created by Git Fusion and changes within it belong to the placeholder user "unknown_git":
```
$ p4 changes save-errno-win32/...
Change 197885 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Fixed typos in the test/functio'
Change 197884 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Save and restore GetLastError()'
Change 197883 on 2021/08/08 by unknown_git@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Added a test case for GitHub is'
Change 197882 on 2021/08/08 by git-fusion-user@git-fusion--temp-perforce-mps-public-8299bb4a-f83a-11eb-a758-00163e71ffc3 'Git Fusion branch management I'
```
When Git Fusion creates a branch like this, it updates the configuration for the repo, so we can track it down there:
```
$ p4 annotate -c //.git-fusion/repos/mps-public/p4gf_config | grep save-errno-win32
197888: git-branch-name = save-errno-win32
...
$ p4 changes //...@197886,197888
Change 197888 on 2021/08/08 by git-fusion-user@git-fusion--perforce-8226a998-f83a-11eb-8966-00163e71ffc3 'Update 'p4gf_config'. '
Change 197887 on 2021/08/08 by git-fusion-user@git-fusion--perforce-8226a998-f83a-11eb-8966-00163e71ffc3 'Git Fusion 'mps-public' copied '
Change 197886 on 2021/08/08 by gdr@git-fusion--temp-perforce-mps-public-82b3ee86-f83a-11eb-b986-00163e71ffc3 'Merge branch 'save-errno-win32''
```
So we can see that this was caused by a merge of a third-party pull request by @gareth-rees :
```
$ p4 describe -s 197886
Change 197886 by gdr@git-fusion--temp-perforce-mps-public-82b3ee86-f83a-11eb-b986-00163e71ffc3 on 2021/08/08 12:19:50
Merge branch 'save-errno-win32' from Filip Strömbäck.
Imported from Git
Author: Gareth Rees [<gdr@garethrees.org>](mailto:gdr@garethrees.org) 1628421098 +0100 (2021-08-08 12:11:38+01:00)
Committer: Gareth Rees [<gdr@garethrees.org>](mailto:gdr@garethrees.org) 1628421098 +0100 (2021-08-08 12:11:38+01:00)
sha1: 86a528d1cdee81ecbf9365e3db7a65b1a3aaa05b
push-state: complete
parent-changes: cb4051bd02a61cc75927bb28aa4d2643a848c3eb=[197876, 197878]/8360988487490952f6edae36b70c1bba00fe6d2f=[197885]
Affected files ...
... //info.ravenbrook.com/project/mps/master/code/protw3.c#31 edit
... //info.ravenbrook.com/project/mps/master/design/thread-manager.txt#26 edit
... //info.ravenbrook.com/project/mps/master/test/function/237.c#1 add
```
This explains the "unknown_git" changes: they were made by @fstromback .
@fstromback's branch was imported into the git repo and can be see at https://github.com/Ravenbrook/mps/tree/save-errno-win32. How'd that happen?
There is no pull request by @fstromback by searching https://github.com/Ravenbrook/mps/pulls?q=is%3Apr+fstromback
This is where we can see flaws in GitHub's system. How do we track this down? Searching for a branch name in the pull requests does not reveal them. (I tested with some known ones.) I just had to make some guesses and found https://github.com/Ravenbrook/mps/pull/62
That pull request resulted in a merge of a branch from a fork of the MPS at https://github.com/fstromback/mps/tree/save-errno-win32 .
Somehow (probably GitHub's default behaviour via the GUI) this created a branch of the same name in Ravenbrook's repo at https://github.com/Ravenbrook/mps/tree/save-errno-win32.
From there I think @gareth-rees followed [his procedure](https://info.ravenbrook.com/mail/2020/09/03/13-02-35/0/) and at step 7 did `git push ravenbrook master "$BRANCH”` at which point Git Fusion created //info.ravenbrook.com/project/mps/save-errno-win32.
|
process
|
merge procedure creates bogus perforce branch locations at some point we gained a bogus top level directory in perforce at info ravenbrook com project mps save errno how d that happen it s something that s very hard to correct how do we prevent it happening again basically this happened because a branch was pushed that did not conform to the branch date topic naming convention git fusion is mapping branch names to perforce directories relative to info ravenbrook com project mps so if a branch is pushed called foo it ll end up at info ravenbrook com project mps foo here s how i found out what happened and my hypothesis of the steps that led to this the branch was created by git fusion and changes within it belong to the placeholder user unknown git changes save errno change on by unknown git git fusion temp perforce mps public fixed typos in the test functio change on by unknown git git fusion temp perforce mps public save and restore getlasterror change on by unknown git git fusion temp perforce mps public added a test case for github is change on by git fusion user git fusion temp perforce mps public git fusion branch management i when git fusion creates a branch like this it updates the configuration for the repo so we can track it down there annotate c git fusion repos mps public config grep save errno git branch name save errno changes change on by git fusion user git fusion perforce update config change on by git fusion user git fusion perforce git fusion mps public copied change on by gdr git fusion temp perforce mps public merge branch save errno so we can see that this was caused by a merge of a third party pull request by gareth rees describe s change by gdr git fusion temp perforce mps public on merge branch save errno from filip strömbäck imported from git author gareth rees mailto gdr garethrees org committer gareth rees mailto gdr garethrees org push state complete parent changes affected files info ravenbrook com project mps master code c edit info ravenbrook com project mps master design thread manager txt edit info ravenbrook com project mps master test function c add this explains the unknown git changes they were made by fstromback fstromback s branch was imported into the git repo and can be see at how d that happen there is no pull request by fstromback by searching this is where we can see flaws in github s system how do we track this down searching for a branch name in the pull requests does not reveal them i tested with some known ones i just had to make some guesses and found that pull request resulted in a merge of a branch from a fork of the mps at somehow probably github s default behaviour via the gui this created a branch of the same name in ravenbrook s repo at from there i think gareth rees followed and at step did git push ravenbrook master branch” at which point git fusion created info ravenbrook com project mps save errno
| 1
|
22,377
| 31,142,282,258
|
IssuesEvent
|
2023-08-16 01:43:56
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: endVideoCapture timeouts
|
process: flaky test topic: flake ❄️ stage: fire watch priority: low stale
|
### Link to dashboard or CircleCI failure
https://dashboard.cypress.io/projects/ypt4pf/runs/38164/test-results/instance/4ec63f4c-8f52-4a2d-9329-6e42bb2e6805/stdout
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/a0a64cc851c8232b94eb0e907f220e08260bda20/packages/server/lib/video_capture.ts#L129
### Analysis
I've seen this pretty frequently. Low priority since it doesn't cause failures
<img width="869" alt="Screen Shot 2022-08-19 at 12 57 34 AM" src="https://user-images.githubusercontent.com/26726429/185572143-161c6d4b-843b-4f08-ad0c-ac691f3d555a.png">
### Cypress Version
10.5.0
### Other
_No response_
|
1.0
|
Flaky test: endVideoCapture timeouts - ### Link to dashboard or CircleCI failure
https://dashboard.cypress.io/projects/ypt4pf/runs/38164/test-results/instance/4ec63f4c-8f52-4a2d-9329-6e42bb2e6805/stdout
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/a0a64cc851c8232b94eb0e907f220e08260bda20/packages/server/lib/video_capture.ts#L129
### Analysis
I've seen this pretty frequently. Low priority since it doesn't cause failures
<img width="869" alt="Screen Shot 2022-08-19 at 12 57 34 AM" src="https://user-images.githubusercontent.com/26726429/185572143-161c6d4b-843b-4f08-ad0c-ac691f3d555a.png">
### Cypress Version
10.5.0
### Other
_No response_
|
process
|
flaky test endvideocapture timeouts link to dashboard or circleci failure link to failing test in github analysis i ve seen this pretty frequently low priority since it doesn t cause failures img width alt screen shot at am src cypress version other no response
| 1
|
383,488
| 26,552,538,105
|
IssuesEvent
|
2023-01-20 09:11:56
|
bazelbuild/intellij
|
https://api.github.com/repos/bazelbuild/intellij
|
closed
|
Documentation: explain how to run tests
|
type: documentation product: IntelliJ type: user support
|
It's easy to build the entire project by selecting Bazel > Build > Compile Project. However, it was unclear how to run all the tests.
It would be good to document this below "Building your project" here: https://ij.bazel.build/docs/bazel-plugin.html#building-your-project
My method was:
1. Run > Edit configurations
1. Click "+"
1. Select "Bazel Command", and fill in:
a. Target expression: //...
b. Bazel command: test
1. OK, then click the "Play" button (Shift+F10)
But I don't know if this is the easiest way.
|
1.0
|
Documentation: explain how to run tests - It's easy to build the entire project by selecting Bazel > Build > Compile Project. However, it was unclear how to run all the tests.
It would be good to document this below "Building your project" here: https://ij.bazel.build/docs/bazel-plugin.html#building-your-project
My method was:
1. Run > Edit configurations
1. Click "+"
1. Select "Bazel Command", and fill in:
a. Target expression: //...
b. Bazel command: test
1. OK, then click the "Play" button (Shift+F10)
But I don't know if this is the easiest way.
|
non_process
|
documentation explain how to run tests it s easy to build the entire project by selecting bazel build compile project however it was unclear how to run all the tests it would be good to document this below building your project here my method was run edit configurations click select bazel command and fill in a target expression b bazel command test ok then click the play button shift but i don t know if this is the easiest way
| 0
|
149,463
| 11,902,287,658
|
IssuesEvent
|
2020-03-30 13:45:43
|
ooni/probe
|
https://api.github.com/repos/ooni/probe
|
opened
|
Run E2E tests in production environment
|
ooni/probe-desktop priority/low testing
|
Currently, the default configuration of Spectron makes it run tests in dev environment. We would like to test the code that is closest to what runs at the user end.
|
1.0
|
Run E2E tests in production environment - Currently, the default configuration of Spectron makes it run tests in dev environment. We would like to test the code that is closest to what runs at the user end.
|
non_process
|
run tests in production environment currently the default configuration of spectron makes it run tests in dev environment we would like to test the code that is closest to what runs at the user end
| 0
|
15,135
| 18,890,025,902
|
IssuesEvent
|
2021-11-15 12:11:40
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
closed
|
Analysis: How to Restructure ProcessAPI and process handling in Altinn App
|
area/process kind/analysis
|
## Description
Based on #2743 we see it is possible to restructure the Process API and Process handling to improve the architecture. The goal for this possible change is
- Make it less complex to add types of task support. Like Payment or Signing.
- Make it less complex to add gateways to add support for more flows
- Reduce code complexity
- Put responsibility where it should be and move business logout out from ProcessController
The new structure will have the following key components
## Principles
- Changes done to process should be triggered by ProcessAPI directly. Not indirectly through other API
### ProcessChange entity
The ProsessChange entity will be an entity that is passed between the different parts described below.
It will contain information about
- Instance - The complete instance object connected to a given process change. That again contains any current process state info
- Status - Information about the result of process change
- Message - A list of messages created by components below used for error messages for API to expose
- Requested process element (task, end event)
- Flow info
Using ProcessChange entity as input and response object from the different parts of the processing logic reduces complexity to add functionality later.
### IProcess and Process
The interface IProcess and the implementation of the Process will have all data about the process like:
- Current State
- Tasks in process
- Gateways in process
- Start and End events in the process
- All the flows in and out of the events, tasks, and gateway
It will expose methods to list information
- GetNextElements()
- GetCurrentTask()
It will expose methods for changing the state of the process
- StartProcess()
- Next(?task)
- CompleteProcess
These methods will be used by process API
It will use the current ProcessService and [BMPNReader](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.Common/Process/BpmnReader.cs) to load the [process.bpmn](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/App/config/process/process.bpmn) in App from the file when the process controller is instantiated.
The process implementation will have all logic needed to understand what kind of tasks and gateway that is defined and what kind of flows are possible. Logic from [ProcessHelper](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.PlatformServices/Helpers/ProcessHelper.cs) will be moved into this if reusable.
### IProcessHandler and ProcessHandler
This will be an interface that will have a custom implementation for Altinn App. This to not have the Process to connected with the business Logic in Altinn App and make unit testing of Process more simple
The methods in the interface would be something like
- HandleProcessStart
- HandleProcessChange
ProcessHandler will be responsible for identifying which type of AltinnTask that should be instantiated when instantiation of correct TaskImplementation class **It would need to supply all services required by the different tasks**
ProcessHandler will also need
and which methods should be run.
As an example. If the flow is going back, the method OnTaskLeave should be run, not OnTaskComplete.
It will use flow information to decide this.
It will also be responsible for the instantiation of Gateway implementation and the code related to that
### TaskImpl
Each Altinn task type will have its separate implementation for some common methods.
- OnTaskEnter - Event that happens when an instance is moved to a specific task
- OnTaskLeave - Events that happen when an instance is leaving a task without completing it
- OnTaskComplete - events that happen when a task is completed
The actual implementation would be responsible for calling AltinnApp and Altinn Platform components similar to what ProcessApi is doing now.
OnTaskComplete would formData and other requirements need to be valid
Some principles would be followed when implementing this
#### Updates to the instance and pushing events/
The updates to the instance of process state should be performed by the current task.
Example
- Updating instance to new taskId or end event
- Publishing instance events about the change
- Publishing events
See technical sequence diagrams below
#### Updates of data
Data changes should be handled by the actual task where data is related. Like when completing a task, the current task should lock data, and when revisiting a task, the target task should be responsible for unlocking data.
### GatewayImpl
Gateways will be used where there are more or two flows from a Task.
- UserControlledGateway - User Control where flow should go. (example: go
- ConfiguredGateWay - Gateway where a user with help of simple config can define where task process should go
- CodedGateway - Requires that app developer code logic
### StartEventImpl
The start event implementation implements the logic that needs to be run when instantiation of an instance process.
### EndEventImpl
### Flow
In the BPMN we need to define what type of flow is.
- CompleteTask - A flow that completes a task
- LeaveTask - A flow that leaves a task without completing task
### Technical Sequence Diagrams
The following sequence diagrams shows how the different parts would work together
#### Start Process during instansiation
This flow describes the following process

This technical sequence flow is different than other process change requests since it is an indirect call from [InstanceController](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.Api/Controllers/InstancesController.cs#L166) during instantiation
The following happens
- End-user system or App Frontend post to instance controller
- Request is validated and authorized by the Instance controller
- Instance controller calls StartsProcess() on Process. This identifies start event.
- ProcessHandler is call with Start()
- ProcessHanlder call startEvent implementation
- Start event implementation calls Instantiation validation
- Start event implementation calls ProcessService to get the next task after
- Start event implementation calls instance service in platform to create Instance
- Instance controller in App Creates any data elements based on multipart request and updates Instance
- InstanceAPI calls Process to start the current task
- Process calls Process Handler to start the task
- ProcessHandler identifies correct TaskImplementation and calls StartTask
- TaskImplementation calls OnStartProcessTask on App
- App creates all required data elements for tasks not created by multipart request
- TaskImpl pushes instance events
- Instance pushes App events to Altinn Events

#### Go from Task 1 to Task 2
This flow describes the following process

In this sequence diagram, the current task on the instance is Task1. User presses next

#### Going back to Task 1
The first sequence diagram is based on the following BPMN process

The sequence diagram is for when task is currently at Task 2 and the user chooses to go back to Task1

#### Optional task
The following sequence diagram describes the processing of the following flow.

The Sequence describes when the user press next when the current task is Task2. Gateway logic will then decide if the user goes to the end event or Task3

## Considerations
It is not feasible to make a process engine that is not someway connected to Altinn. This makes it less reusable for others. OK?
Can we continue using BPMNReader? The assumption is yes.
Can we move functionality from processHelper to process? The assumption is yes.
To keep this change possible to review we should separate the implementation and pull requests into smaller parts.
See suggested development parts. This could be separated in separate issues for the Zenhub board.
### Ops requirements
> Are there any requirements for monitoring? What is being built and what could go wrong?
> Are there any requirements related to backup?
## Acceptance criteria
- [ ] The change should be backward compatible
- [ ] Logic is moved out from ProcessApi and structured as described above
- [ ] App supports user-controlled gateways
- [ ] API exposes possible gateway options to frontend
- [ ] ProcessAPI interface does not change
- [ ] The way data is updated does not change
## Specification tasks
- [ ] Development tasks are defined
### Development Task
Part 1
- [ ] Define IProcess engine interface with response models
- [ ] Implement Processengine
- [ ] Define IProcessHandler interface
- [ ] Implement unit tests for process engine
Part 2
- [ ] Implement IProcessHandler for Altinn App
- [ ] Implement TaskImp for Data
- [ ] Implement TaskImp for Confirm
- [ ] Implement TaskImpl for Feedback
- [ ] Create unit test that that verify flows that include Altinn app and the different current task types
Part 4
- [ ] Update BPMN reader to support gateways
- [ ] Create Gateway impl for coded gateway
- [ ] Implement Unit tests
Part 5
- [ ] Update Process API to use new Process structure
- [ ] Verify current unit test that covers process API to make sure they are good enough to cover the changes
- [ ] Add unit test for ProcessAPi that coves gateway scenarios
Part 6
- [ ] Update InstanceAPI to use new Process structure
- [ ] Verify unit tests
## Test
- [ ] Regression test standard process standard
- [ ] Test process with gateway
## Definition of done
Verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing.
- [ ] Documentation is updated (if relevant)
- [ ] Technical documentation (docs.altinn.studio)
- [ ] User documentation (altinn.github.io/docs)
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
|
1.0
|
Analysis: How to Restructure ProcessAPI and process handling in Altinn App - ## Description
Based on #2743 we see it is possible to restructure the Process API and Process handling to improve the architecture. The goal for this possible change is
- Make it less complex to add types of task support. Like Payment or Signing.
- Make it less complex to add gateways to add support for more flows
- Reduce code complexity
- Put responsibility where it should be and move business logout out from ProcessController
The new structure will have the following key components
## Principles
- Changes done to process should be triggered by ProcessAPI directly. Not indirectly through other API
### ProcessChange entity
The ProsessChange entity will be an entity that is passed between the different parts described below.
It will contain information about
- Instance - The complete instance object connected to a given process change. That again contains any current process state info
- Status - Information about the result of process change
- Message - A list of messages created by components below used for error messages for API to expose
- Requested process element (task, end event)
- Flow info
Using ProcessChange entity as input and response object from the different parts of the processing logic reduces complexity to add functionality later.
### IProcess and Process
The interface IProcess and the implementation of the Process will have all data about the process like:
- Current State
- Tasks in process
- Gateways in process
- Start and End events in the process
- All the flows in and out of the events, tasks, and gateway
It will expose methods to list information
- GetNextElements()
- GetCurrentTask()
It will expose methods for changing the state of the process
- StartProcess()
- Next(?task)
- CompleteProcess
These methods will be used by process API
It will use the current ProcessService and [BMPNReader](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.Common/Process/BpmnReader.cs) to load the [process.bpmn](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/App/config/process/process.bpmn) in App from the file when the process controller is instantiated.
The process implementation will have all logic needed to understand what kind of tasks and gateway that is defined and what kind of flows are possible. Logic from [ProcessHelper](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.PlatformServices/Helpers/ProcessHelper.cs) will be moved into this if reusable.
### IProcessHandler and ProcessHandler
This will be an interface that will have a custom implementation for Altinn App. This to not have the Process to connected with the business Logic in Altinn App and make unit testing of Process more simple
The methods in the interface would be something like
- HandleProcessStart
- HandleProcessChange
ProcessHandler will be responsible for identifying which type of AltinnTask that should be instantiated when instantiation of correct TaskImplementation class **It would need to supply all services required by the different tasks**
ProcessHandler will also need
and which methods should be run.
As an example. If the flow is going back, the method OnTaskLeave should be run, not OnTaskComplete.
It will use flow information to decide this.
It will also be responsible for the instantiation of Gateway implementation and the code related to that
### TaskImpl
Each Altinn task type will have its separate implementation for some common methods.
- OnTaskEnter - Event that happens when an instance is moved to a specific task
- OnTaskLeave - Events that happen when an instance is leaving a task without completing it
- OnTaskComplete - events that happen when a task is completed
The actual implementation would be responsible for calling AltinnApp and Altinn Platform components similar to what ProcessApi is doing now.
OnTaskComplete would formData and other requirements need to be valid
Some principles would be followed when implementing this
#### Updates to the instance and pushing events/
The updates to the instance of process state should be performed by the current task.
Example
- Updating instance to new taskId or end event
- Publishing instance events about the change
- Publishing events
See technical sequence diagrams below
#### Updates of data
Data changes should be handled by the actual task where data is related. Like when completing a task, the current task should lock data, and when revisiting a task, the target task should be responsible for unlocking data.
### GatewayImpl
Gateways will be used where there are more or two flows from a Task.
- UserControlledGateway - User Control where flow should go. (example: go
- ConfiguredGateWay - Gateway where a user with help of simple config can define where task process should go
- CodedGateway - Requires that app developer code logic
### StartEventImpl
The start event implementation implements the logic that needs to be run when instantiation of an instance process.
### EndEventImpl
### Flow
In the BPMN we need to define what type of flow is.
- CompleteTask - A flow that completes a task
- LeaveTask - A flow that leaves a task without completing task
### Technical Sequence Diagrams
The following sequence diagrams shows how the different parts would work together
#### Start Process during instansiation
This flow describes the following process

This technical sequence flow is different than other process change requests since it is an indirect call from [InstanceController](https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/Altinn.App.Api/Controllers/InstancesController.cs#L166) during instantiation
The following happens
- End-user system or App Frontend post to instance controller
- Request is validated and authorized by the Instance controller
- Instance controller calls StartsProcess() on Process. This identifies start event.
- ProcessHandler is call with Start()
- ProcessHanlder call startEvent implementation
- Start event implementation calls Instantiation validation
- Start event implementation calls ProcessService to get the next task after
- Start event implementation calls instance service in platform to create Instance
- Instance controller in App Creates any data elements based on multipart request and updates Instance
- InstanceAPI calls Process to start the current task
- Process calls Process Handler to start the task
- ProcessHandler identifies correct TaskImplementation and calls StartTask
- TaskImplementation calls OnStartProcessTask on App
- App creates all required data elements for tasks not created by multipart request
- TaskImpl pushes instance events
- Instance pushes App events to Altinn Events

#### Go from Task 1 to Task 2
This flow describes the following process

In this sequence diagram, the current task on the instance is Task1. User presses next

#### Going back to Task 1
The first sequence diagram is based on the following BPMN process

The sequence diagram is for when task is currently at Task 2 and the user chooses to go back to Task1

#### Optional task
The following sequence diagram describes the processing of the following flow.

The Sequence describes when the user press next when the current task is Task2. Gateway logic will then decide if the user goes to the end event or Task3

## Considerations
It is not feasible to make a process engine that is not someway connected to Altinn. This makes it less reusable for others. OK?
Can we continue using BPMNReader? The assumption is yes.
Can we move functionality from processHelper to process? The assumption is yes.
To keep this change possible to review we should separate the implementation and pull requests into smaller parts.
See suggested development parts. This could be separated in separate issues for the Zenhub board.
### Ops requirements
> Are there any requirements for monitoring? What is being built and what could go wrong?
> Are there any requirements related to backup?
## Acceptance criteria
- [ ] The change should be backward compatible
- [ ] Logic is moved out from ProcessApi and structured as described above
- [ ] App supports user-controlled gateways
- [ ] API exposes possible gateway options to frontend
- [ ] ProcessAPI interface does not change
- [ ] The way data is updated does not change
## Specification tasks
- [ ] Development tasks are defined
### Development Task
Part 1
- [ ] Define IProcess engine interface with response models
- [ ] Implement Processengine
- [ ] Define IProcessHandler interface
- [ ] Implement unit tests for process engine
Part 2
- [ ] Implement IProcessHandler for Altinn App
- [ ] Implement TaskImp for Data
- [ ] Implement TaskImp for Confirm
- [ ] Implement TaskImpl for Feedback
- [ ] Create unit test that that verify flows that include Altinn app and the different current task types
Part 4
- [ ] Update BPMN reader to support gateways
- [ ] Create Gateway impl for coded gateway
- [ ] Implement Unit tests
Part 5
- [ ] Update Process API to use new Process structure
- [ ] Verify current unit test that covers process API to make sure they are good enough to cover the changes
- [ ] Add unit test for ProcessAPi that coves gateway scenarios
Part 6
- [ ] Update InstanceAPI to use new Process structure
- [ ] Verify unit tests
## Test
- [ ] Regression test standard process standard
- [ ] Test process with gateway
## Definition of done
Verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing.
- [ ] Documentation is updated (if relevant)
- [ ] Technical documentation (docs.altinn.studio)
- [ ] User documentation (altinn.github.io/docs)
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
|
process
|
analysis how to restructure processapi and process handling in altinn app description based on we see it is possible to restructure the process api and process handling to improve the architecture the goal for this possible change is make it less complex to add types of task support like payment or signing make it less complex to add gateways to add support for more flows reduce code complexity put responsibility where it should be and move business logout out from processcontroller the new structure will have the following key components principles changes done to process should be triggered by processapi directly not indirectly through other api processchange entity the prosesschange entity will be an entity that is passed between the different parts described below it will contain information about instance the complete instance object connected to a given process change that again contains any current process state info status information about the result of process change message a list of messages created by components below used for error messages for api to expose requested process element task end event flow info using processchange entity as input and response object from the different parts of the processing logic reduces complexity to add functionality later iprocess and process the interface iprocess and the implementation of the process will have all data about the process like current state tasks in process gateways in process start and end events in the process all the flows in and out of the events tasks and gateway it will expose methods to list information getnextelements getcurrenttask it will expose methods for changing the state of the process startprocess next task completeprocess these methods will be used by process api it will use the current processservice and to load the in app from the file when the process controller is instantiated the process implementation will have all logic needed to understand what kind of tasks and gateway that is defined and what kind of flows are possible logic from will be moved into this if reusable iprocesshandler and processhandler this will be an interface that will have a custom implementation for altinn app this to not have the process to connected with the business logic in altinn app and make unit testing of process more simple the methods in the interface would be something like handleprocessstart handleprocesschange processhandler will be responsible for identifying which type of altinntask that should be instantiated when instantiation of correct taskimplementation class it would need to supply all services required by the different tasks processhandler will also need and which methods should be run as an example if the flow is going back the method ontaskleave should be run not ontaskcomplete it will use flow information to decide this it will also be responsible for the instantiation of gateway implementation and the code related to that taskimpl each altinn task type will have its separate implementation for some common methods ontaskenter event that happens when an instance is moved to a specific task ontaskleave events that happen when an instance is leaving a task without completing it ontaskcomplete events that happen when a task is completed the actual implementation would be responsible for calling altinnapp and altinn platform components similar to what processapi is doing now ontaskcomplete would formdata and other requirements need to be valid some principles would be followed when implementing this updates to the instance and pushing events the updates to the instance of process state should be performed by the current task example updating instance to new taskid or end event publishing instance events about the change publishing events see technical sequence diagrams below updates of data data changes should be handled by the actual task where data is related like when completing a task the current task should lock data and when revisiting a task the target task should be responsible for unlocking data gatewayimpl gateways will be used where there are more or two flows from a task usercontrolledgateway user control where flow should go example go configuredgateway gateway where a user with help of simple config can define where task process should go codedgateway requires that app developer code logic starteventimpl the start event implementation implements the logic that needs to be run when instantiation of an instance process endeventimpl flow in the bpmn we need to define what type of flow is completetask a flow that completes a task leavetask a flow that leaves a task without completing task technical sequence diagrams the following sequence diagrams shows how the different parts would work together start process during instansiation this flow describes the following process this technical sequence flow is different than other process change requests since it is an indirect call from during instantiation the following happens end user system or app frontend post to instance controller request is validated and authorized by the instance controller instance controller calls startsprocess on process this identifies start event processhandler is call with start processhanlder call startevent implementation start event implementation calls instantiation validation start event implementation calls processservice to get the next task after start event implementation calls instance service in platform to create instance instance controller in app creates any data elements based on multipart request and updates instance instanceapi calls process to start the current task process calls process handler to start the task processhandler identifies correct taskimplementation and calls starttask taskimplementation calls onstartprocesstask on app app creates all required data elements for tasks not created by multipart request taskimpl pushes instance events instance pushes app events to altinn events go from task to task this flow describes the following process in this sequence diagram the current task on the instance is user presses next going back to task the first sequence diagram is based on the following bpmn process the sequence diagram is for when task is currently at task and the user chooses to go back to optional task the following sequence diagram describes the processing of the following flow the sequence describes when the user press next when the current task is gateway logic will then decide if the user goes to the end event or considerations it is not feasible to make a process engine that is not someway connected to altinn this makes it less reusable for others ok can we continue using bpmnreader the assumption is yes can we move functionality from processhelper to process the assumption is yes to keep this change possible to review we should separate the implementation and pull requests into smaller parts see suggested development parts this could be separated in separate issues for the zenhub board ops requirements are there any requirements for monitoring what is being built and what could go wrong are there any requirements related to backup acceptance criteria the change should be backward compatible logic is moved out from processapi and structured as described above app supports user controlled gateways api exposes possible gateway options to frontend processapi interface does not change the way data is updated does not change specification tasks development tasks are defined development task part define iprocess engine interface with response models implement processengine define iprocesshandler interface implement unit tests for process engine part implement iprocesshandler for altinn app implement taskimp for data implement taskimp for confirm implement taskimpl for feedback create unit test that that verify flows that include altinn app and the different current task types part update bpmn reader to support gateways create gateway impl for coded gateway implement unit tests part update process api to use new process structure verify current unit test that covers process api to make sure they are good enough to cover the changes add unit test for processapi that coves gateway scenarios part update instanceapi to use new process structure verify unit tests test regression test standard process standard test process with gateway definition of done verify that this issue meets only for project members before closing documentation is updated if relevant technical documentation docs altinn studio user documentation altinn github io docs qa manual test is complete if relevant automated test is implemented if relevant all tasks in this userstory are closed i e remaining tasks are moved to other user stories or marked obsolete
| 1
|
48,053
| 5,932,928,172
|
IssuesEvent
|
2017-05-24 10:35:23
|
EenmaalAndermaal/EenmaalAndermaal
|
https://api.github.com/repos/EenmaalAndermaal/EenmaalAndermaal
|
closed
|
verkoper upgrade is kapot
|
bug prioriteit: 1 tester: Thijs
|
# Omschrijving
Kolomnaam was gebruikersnaam is omgezet naar gebruiker.
Wordt gebruikt in bestanden:
- upgrade.php
|
1.0
|
verkoper upgrade is kapot - # Omschrijving
Kolomnaam was gebruikersnaam is omgezet naar gebruiker.
Wordt gebruikt in bestanden:
- upgrade.php
|
non_process
|
verkoper upgrade is kapot omschrijving kolomnaam was gebruikersnaam is omgezet naar gebruiker wordt gebruikt in bestanden upgrade php
| 0
|
14,777
| 25,730,789,635
|
IssuesEvent
|
2022-12-07 20:08:32
|
Croquembouche/pyWhat-2022
|
https://api.github.com/repos/Croquembouche/pyWhat-2022
|
closed
|
NFR 5: Portability
|
Product Backlog Non-Functional Requirements
|
|NFR 5: Portability|
|----------------------|
| **Estimate:** 8 |
| **Priority:** Must have |
|**Story**: As a user, I would like to use pywhat on all python supported platforms.|
| **Acceptance criteria:** |
|1.The application can be used and returns the correct results on all python supported platforms.|
|
1.0
|
NFR 5: Portability - |NFR 5: Portability|
|----------------------|
| **Estimate:** 8 |
| **Priority:** Must have |
|**Story**: As a user, I would like to use pywhat on all python supported platforms.|
| **Acceptance criteria:** |
|1.The application can be used and returns the correct results on all python supported platforms.|
|
non_process
|
nfr portability nfr portability estimate priority must have story as a user i would like to use pywhat on all python supported platforms acceptance criteria the application can be used and returns the correct results on all python supported platforms
| 0
|
17,697
| 23,547,286,218
|
IssuesEvent
|
2022-08-21 10:00:18
|
Battle-s/battle-school-backend
|
https://api.github.com/repos/Battle-s/battle-school-backend
|
closed
|
[FEAT] Status Enum common에 생성
|
feature :computer: processing :hourglass_flowing_sand:
|
## 설명
> Status 컬럼을 쓰는 곳이 몇개 있어서 common에 생성합니다.
## 체크사항
> 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다.
- [ ] todo1
- [ ] todo2
- [ ] todo3
## 참고자료
> 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다.
## 관련 논의
> 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다.
|
1.0
|
[FEAT] Status Enum common에 생성 - ## 설명
> Status 컬럼을 쓰는 곳이 몇개 있어서 common에 생성합니다.
## 체크사항
> 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다.
- [ ] todo1
- [ ] todo2
- [ ] todo3
## 참고자료
> 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다.
## 관련 논의
> 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다.
|
process
|
status enum common에 생성 설명 status 컬럼을 쓰는 곳이 몇개 있어서 common에 생성합니다 체크사항 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다 참고자료 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다 관련 논의 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다
| 1
|
221,676
| 24,656,926,290
|
IssuesEvent
|
2022-10-18 01:02:59
|
btmluiz/prometeuspbx
|
https://api.github.com/repos/btmluiz/prometeuspbx
|
opened
|
Django-4.1-py3-none-any.whl: 1 vulnerabilities (highest severity is: 5.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-4.1-py3-none-any.whl</b></p></summary>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl">https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl</a></p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-41323](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-41323) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Django-4.1-py3-none-any.whl | Direct | Django - 3.2.16,4.0.8,4.1.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-41323</summary>
### Vulnerable Library - <b>Django-4.1-py3-none-any.whl</b></p>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl">https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **Django-4.1-py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Django 3.2 before 3.2.16, 4.0 before 4.0.8, and 4.1 before 4.1.2, internationalized URLs were subject to a potential denial of service attack via the locale parameter, which is treated as a regular expression.
<p>Publish Date: 2022-10-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-41323>CVE-2022-41323</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-41323">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-41323</a></p>
<p>Release Date: 2022-10-16</p>
<p>Fix Resolution: Django - 3.2.16,4.0.8,4.1.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
Django-4.1-py3-none-any.whl: 1 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-4.1-py3-none-any.whl</b></p></summary>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl">https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl</a></p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-41323](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-41323) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Django-4.1-py3-none-any.whl | Direct | Django - 3.2.16,4.0.8,4.1.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-41323</summary>
### Vulnerable Library - <b>Django-4.1-py3-none-any.whl</b></p>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl">https://files.pythonhosted.org/packages/9b/41/e1e7d6ecc3bc76681dfdc6b373566822bc2aab96fa3eceaaed70accc28b6/Django-4.1-py3-none-any.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **Django-4.1-py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Django 3.2 before 3.2.16, 4.0 before 4.0.8, and 4.1 before 4.1.2, internationalized URLs were subject to a potential denial of service attack via the locale parameter, which is treated as a regular expression.
<p>Publish Date: 2022-10-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-41323>CVE-2022-41323</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-41323">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-41323</a></p>
<p>Release Date: 2022-10-16</p>
<p>Fix Resolution: Django - 3.2.16,4.0.8,4.1.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
django none any whl vulnerabilities highest severity is vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium django none any whl direct django details cve vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href dependency hierarchy x django none any whl vulnerable library found in base branch master vulnerability details in django before before and before internationalized urls were subject to a potential denial of service attack via the locale parameter which is treated as a regular expression publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with mend
| 0
|
6,115
| 8,973,911,999
|
IssuesEvent
|
2019-01-29 22:24:24
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Disable `install` target when using external projects.
|
type: process
|
When using external projects `make install` should be disabled: the dependencies are not installed in the system, and installing `google-cloud-cpp` would fail.
|
1.0
|
Disable `install` target when using external projects. - When using external projects `make install` should be disabled: the dependencies are not installed in the system, and installing `google-cloud-cpp` would fail.
|
process
|
disable install target when using external projects when using external projects make install should be disabled the dependencies are not installed in the system and installing google cloud cpp would fail
| 1
|
6,418
| 9,516,133,794
|
IssuesEvent
|
2019-04-26 08:03:31
|
StefanPahlplatz/proep-backend
|
https://api.github.com/repos/StefanPahlplatz/proep-backend
|
closed
|
Create a list of user stories
|
process
|
- [ ] Create a word document with issues
- [ ] Enter the issues into GitHub
|
1.0
|
Create a list of user stories - - [ ] Create a word document with issues
- [ ] Enter the issues into GitHub
|
process
|
create a list of user stories create a word document with issues enter the issues into github
| 1
|
7,578
| 10,687,100,460
|
IssuesEvent
|
2019-10-22 15:31:19
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
GO:0080185 effector-dependent induction by symbiont of host immune response
|
multi-species process
|
GO:0080185 effector-dependent induction by symbiont of host immune response
GO:0080185 JSON
effector-dependent induction by symbiont of host immune response
Biological Process
Definition (GO:0080185 GONUTS page)
Any process that involves recognition of an effector, and by which a symbiont activates, maintains or increases the frequency, rate or extent of the immune response of the host organism; the immune response is any immune system process that functions in the calibrated response of an organism to a potential internal or invasive threat. The host is defined as the larger of the organisms involved in a symbiotic interaction. Effectors are proteins secreted into the host cell by pathogenic microbes, presumably to alter host immune response signaling. The best characterized effectors are bacterial effectors delivered into the host cell by type III secretion system (TTSS). Effector-triggered immunity (ETI) involves the direct or indirect recognition of an effector protein by the host (for example through plant resistance or R proteins) and subsequent activation of host immune response. PMID:16497589
we need to discuss what this can be used for.
I think you would only annotate this process from a pathogen perspective if it was intentional?
I.e necrotrophs.
From the pathogen perspective we would . be annotatin immune aviodence.
On Friday . I'd like to look at gthe parentage of this.
|
1.0
|
GO:0080185 effector-dependent induction by symbiont of host immune response - GO:0080185 effector-dependent induction by symbiont of host immune response
GO:0080185 JSON
effector-dependent induction by symbiont of host immune response
Biological Process
Definition (GO:0080185 GONUTS page)
Any process that involves recognition of an effector, and by which a symbiont activates, maintains or increases the frequency, rate or extent of the immune response of the host organism; the immune response is any immune system process that functions in the calibrated response of an organism to a potential internal or invasive threat. The host is defined as the larger of the organisms involved in a symbiotic interaction. Effectors are proteins secreted into the host cell by pathogenic microbes, presumably to alter host immune response signaling. The best characterized effectors are bacterial effectors delivered into the host cell by type III secretion system (TTSS). Effector-triggered immunity (ETI) involves the direct or indirect recognition of an effector protein by the host (for example through plant resistance or R proteins) and subsequent activation of host immune response. PMID:16497589
we need to discuss what this can be used for.
I think you would only annotate this process from a pathogen perspective if it was intentional?
I.e necrotrophs.
From the pathogen perspective we would . be annotatin immune aviodence.
On Friday . I'd like to look at gthe parentage of this.
|
process
|
go effector dependent induction by symbiont of host immune response go effector dependent induction by symbiont of host immune response go json effector dependent induction by symbiont of host immune response biological process definition go gonuts page any process that involves recognition of an effector and by which a symbiont activates maintains or increases the frequency rate or extent of the immune response of the host organism the immune response is any immune system process that functions in the calibrated response of an organism to a potential internal or invasive threat the host is defined as the larger of the organisms involved in a symbiotic interaction effectors are proteins secreted into the host cell by pathogenic microbes presumably to alter host immune response signaling the best characterized effectors are bacterial effectors delivered into the host cell by type iii secretion system ttss effector triggered immunity eti involves the direct or indirect recognition of an effector protein by the host for example through plant resistance or r proteins and subsequent activation of host immune response pmid we need to discuss what this can be used for i think you would only annotate this process from a pathogen perspective if it was intentional i e necrotrophs from the pathogen perspective we would be annotatin immune aviodence on friday i d like to look at gthe parentage of this
| 1
|
147,143
| 5,634,331,947
|
IssuesEvent
|
2017-04-05 21:04:09
|
vmware/vic
|
https://api.github.com/repos/vmware/vic
|
closed
|
Harbor's Password policy should be configurable for organizations
|
area/security priority/high product/harbor
|
According to Security testing criteria,
`Password policy should be enforced and should be configurable for organizations` ,
But Harbor only have one password policy and not configurable.
|
1.0
|
Harbor's Password policy should be configurable for organizations - According to Security testing criteria,
`Password policy should be enforced and should be configurable for organizations` ,
But Harbor only have one password policy and not configurable.
|
non_process
|
harbor s password policy should be configurable for organizations according to security testing criteria password policy should be enforced and should be configurable for organizations but harbor only have one password policy and not configurable
| 0
|
121,586
| 15,983,043,445
|
IssuesEvent
|
2021-04-18 07:23:28
|
icat-lab/g4n
|
https://api.github.com/repos/icat-lab/g4n
|
closed
|
slackの位置づけと利用方法
|
YFseminar design documentation
|
学部生向けにあまり多数のツールを一度に用いるのはNGだと思いますが、LINEでやりとりではなくSlackは使わせておきたい気がします。
Slackは別にWorkspaceを立てるのではなく、icat_labに追加でチャネルを設置することを想定。
春学期は、当初は1つのチャネルのみ設定。必要に応じて追加。
秋学期は、全体チャネルに加えて、グループごとにチャネルを作成。グループワークで用いる。
という感じを想定します。
[MBAのセットアップドキュメントから一部簡略化したセットアップドキュメント](https://github.com/icat-lab/icat_lab/blob/master/basic_seminar_undergrad/setup.md#slack)を用意しています。こちらに確定事項を反映させていきます。
|
1.0
|
slackの位置づけと利用方法 - 学部生向けにあまり多数のツールを一度に用いるのはNGだと思いますが、LINEでやりとりではなくSlackは使わせておきたい気がします。
Slackは別にWorkspaceを立てるのではなく、icat_labに追加でチャネルを設置することを想定。
春学期は、当初は1つのチャネルのみ設定。必要に応じて追加。
秋学期は、全体チャネルに加えて、グループごとにチャネルを作成。グループワークで用いる。
という感じを想定します。
[MBAのセットアップドキュメントから一部簡略化したセットアップドキュメント](https://github.com/icat-lab/icat_lab/blob/master/basic_seminar_undergrad/setup.md#slack)を用意しています。こちらに確定事項を反映させていきます。
|
non_process
|
slackの位置づけと利用方法 学部生向けにあまり多数のツールを一度に用いるのはngだと思いますが、lineでやりとりではなくslackは使わせておきたい気がします。 slackは別にworkspaceを立てるのではなく、icat labに追加でチャネルを設置することを想定。 春学期は、 。必要に応じて追加。 秋学期は、全体チャネルに加えて、グループごとにチャネルを作成。グループワークで用いる。 という感じを想定します。
| 0
|
20,126
| 26,660,633,885
|
IssuesEvent
|
2023-01-25 20:46:01
|
openxla/stablehlo
|
https://api.github.com/repos/openxla/stablehlo
|
closed
|
Consider removing hlo_test_infer.get_return_type_components.
|
Type inference Process
|
Remove all tests that depend on hlo_test_infer.get_return_type_components because they can be now covered using hlo_test_infer.get_return_types instead.
|
1.0
|
Consider removing hlo_test_infer.get_return_type_components. - Remove all tests that depend on hlo_test_infer.get_return_type_components because they can be now covered using hlo_test_infer.get_return_types instead.
|
process
|
consider removing hlo test infer get return type components remove all tests that depend on hlo test infer get return type components because they can be now covered using hlo test infer get return types instead
| 1
|
446,614
| 31,551,042,006
|
IssuesEvent
|
2023-09-02 04:06:08
|
OpenIMSDK/Open-IM-Server
|
https://api.github.com/repos/OpenIMSDK/Open-IM-Server
|
opened
|
Optimization and Refactoring of Docker and Docker Compose Network Modes
|
kind/documentation enhancement kind/feature
|
# Optimization and Refactoring of Docker and Docker Compose Network Modes
**Description:**
#### Background:
Our current Docker and Docker Compose setups use a bridge network mode. However, there's a need to optimize and refactor the network modes for better scalability, maintainability, and performance.
#### Issue:
Based on our existing `docker-compose.yml` file (see attached), there are areas that can be potentially improved in terms of networking:
1. **Use of Bridge Mode:** While bridge mode is a default and often suitable for many applications, there might be scenarios where host or overlay networking modes may offer better performance or more suitable isolation.
2. **Network Addressing:** Hardcoding IP addresses (e.g., `${MYSQL_NETWORK_ADDRESS}`, `${MONGO_NETWORK_ADDRESS}`, etc.) may not be scalable. We need a more dynamic way of handling this.
3. **Dependency Management:** While `depends_on` is used, we might want to look into the use of `wait-for-it` or similar scripts to ensure services are fully up and ready before dependants start.
4. **Consistency:** Ensure that the network configurations are consistent across all services. For instance, all services should be part of the necessary networks, and network mode should be uniform unless there's a valid reason for deviations.
#### Proposed Action:
1. Evaluate the current bridge mode against our application's needs. Consider if host, overlay, or macvlan modes might be more appropriate.
2. Investigate using Docker's native DNS for service discovery rather than hardcoding IP addresses.
3. Implement `wait-for-it` or similar scripts for better management of service dependencies.
4. Ensure all services have consistent network configurations, including necessary settings for IPv4 and IPv6 (if applicable).
5. Update and refactor the `docker-compose.yml` file based on the findings and proposed changes.
## Content
```
#fixme Clone openIM Server project before using docker-compose,project address:https://github.com/OpenIMSDK/Open-IM-Server.git
version: '3'
networks:
openim-server:
driver: bridge
ipam:
driver: default
config:
- subnet: '${DOCKER_BRIDGE_SUBNET}'
gateway: '${DOCKER_BRIDGE_GATEWAY}'
services:
mysql:
image: mysql:5.7
ports:
- "${MYSQL_PORT}:3306"
container_name: mysql
volumes:
- "${DATA_DIR}/components/mysql/data:/var/lib/mysql"
- "/etc/localtime:/etc/localtime"
environment:
MYSQL_ROOT_PASSWORD: "${MYSQL_PASSWORD}"
restart: always
networks:
openim-server:
ipv4_address: ${MYSQL_NETWORK_ADDRESS}
mongodb:
image: mongo:6.0.2
ports:
- "${MONGO_PORT}:27017"
container_name: mongo
command: --wiredTigerCacheSizeGB 1 --auth
volumes:
- "${DATA_DIR}/components/mongodb/data/db:/data/db"
- "${DATA_DIR}/components/mongodb/data/logs:/data/logs"
- "${DATA_DIR}/components/mongodb/data/conf:/etc/mongo"
- "./openim-server/${SERVER_BRANCH}/scripts/mongo-init.sh:/docker-entrypoint-initdb.d/mongo-init.sh:ro"
environment:
- TZ=Asia/Shanghai
- wiredTigerCacheSizeGB=1
- MONGO_INITDB_ROOT_USERNAME=${MONGO_USERNAME}
- MONGO_INITDB_ROOT_PASSWORD=${MONGO_PASSWORD}
- MONGO_INITDB_DATABASE=${MONGO_DATABASE}
restart: always
networks:
openim-server:
ipv4_address: ${MONGO_NETWORK_ADDRESS}
redis:
image: redis:7.0.0
container_name: redis
ports:
- "${REDIS_PORT}:6379"
volumes:
- "${DATA_DIR}/components/redis/data:/data"
- "${DATA_DIR}/components/redis/config/redis.conf:/usr/local/redis/config/redis.conf"
environment:
TZ: Asia/Shanghai
restart: always
sysctls:
net.core.somaxconn: 1024
command: redis-server --requirepass ${REDIS_PASSWORD} --appendonly yes
networks:
openim-server:
ipv4_address: ${REDIS_NETWORK_ADDRESS}
zookeeper:
image: bitnami/zookeeper:3.8
container_name: zookeeper
ports:
- target: 2181
published: "${ZOOKEEPER_PORT}"
protocol: tcp
volumes:
- "/etc/localtime:/etc/localtime"
environment:
- ALLOW_ANONYMOUS_LOGIN=yes
- TZ="Asia/Shanghai"
restart: always
networks:
openim-server:
ipv4_address: ${ZOOKEEPER_NETWORK_ADDRESS}
kafka:
image: 'bitnami/kafka:latest'
container_name: kafka
restart: always
ports:
- target: 9092
published: "${KAFKA_PORT}"
protocol: tcp
volumes:
- ./openim-server/${SERVER_BRANCH}/scripts/create_topic.sh:/opt/bitnami/kafka/create_topic.sh
# - ${DATA_DIR}/components/kafka/data:/bitnami/kafka/data
command: >
bash -c "
/opt/bitnami/scripts/kafka/run.sh & sleep 30; /opt/bitnami/kafka/create_topic.sh; wait
"
environment:
- KAFKA_CFG_NODE_ID=0
- TZ=Asia/Shanghai
- KAFKA_CFG_PROCESS_ROLES=controller,broker
- KAFKA_CFG_LISTENERS=PLAINTEXT://:9092,CONTROLLER://:9093
- KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP=CONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT
- KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@kafka:9093
- KAFKA_CFG_CONTROLLER_LISTENER_NAMES=CONTROLLER
networks:
openim-server:
ipv4_address: ${KAFKA_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
minio:
image: minio/minio
ports:
- "${MINIO_PORT}:9000"
- "9090:9090"
container_name: minio
volumes:
- "${DATA_DIR}/components/mnt/data:/data"
- "${DATA_DIR}/components/mnt/config:/root/.minio"
environment:
MINIO_ROOT_USER: "${MINIO_ACCESS_KEY}"
MINIO_ROOT_PASSWORD: "${MINIO_SECRET_KEY}"
restart: always
command: minio server /data --console-address ':9090'
networks:
openim-server:
ipv4_address: ${MINIO_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-web:
image: ghcr.io/openimsdk/openim-web:latest
container_name: openim-web
environment:
- OPENIM_WEB_DIST_PATH=${OPENIM_WEB_DIST_PATH}
- OPENIM_WEB_PORT=${OPENIM_WEB_PORT}
restart: always
ports:
- "${OPENIM_WEB_PORT}:11001"
networks:
openim-server:
ipv4_address: ${OPENIM_WEB_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-server:
# image: ghcr.io/openimsdk/openim-server:${SERVER_BRANCH}
image: registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-server:${SERVER_BRANCH}
# image: openim/openim-server:${SERVER_BRANCH}
# build: .
container_name: openim-server
ports:
- ${OPENIM_WS_PORT}:10001
- ${API_OPENIM_PORT}:10002
healthcheck:
test: ["CMD", "/openim/openim-server/scripts/check-all.sh"]
interval: 300s
timeout: 10s
retries: 5
volumes:
- ./openim-server/logs:/openim/openim-server/logs
- ./openim-server/_output:/openim/openim-server/_output
- ./openim-server/${SERVER_BRANCH}/config:/openim/openim-server/config
- ./openim-server/${SERVER_BRANCH}/scripts:/openim/openim-server/scripts
restart: always
depends_on:
- kafka
- mysql
- mongodb
- redis
- minio
logging:
driver: json-file
options:
max-size: "1g"
max-file: "2"
networks:
openim-server:
ipv4_address: ${OPENIM_SERVER_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-chat:
# image: ghcr.io/openimsdk/openim-chat:${CHAT_BRANCH}
image: registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-chat:${CHAT_BRANCH}
# image: ghcr.io/openimsdk/openim-chat:${CHAT_BRANCH}
container_name: openim-chat
healthcheck:
test: ["CMD", "/openim/openim-chat/scripts/check_all.sh"]
interval: 300s
timeout: 10s
retries: 5
ports:
- ${OPENIM_CHAT_API_PORT}:10008
- ${OPENIM_ADMIN_API_PORT}:10009
volumes:
- ./openim-chat/logs:/openim/openim-chat/logs
- ./openim-chat/_output:/openim/openim-chat/_output
- ./openim-chat/${CHAT_BRANCH}/config:/openim/openim-chat/config
- ./openim-chat/${CHAT_BRANCH}/scripts:/openim/openim-chat/scripts
restart: always
user: root:root
depends_on:
- mysql
- mongodb
- redis
- minio
- openim-server
logging:
driver: json-file
options:
max-size: "1g"
max-file: "2"
networks:
openim-server:
ipv4_address: ${OPENIM_CHAT_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
prometheus:
image: prom/prometheus
volumes:
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/prometheus-compose.yml:/etc/prometheus/prometheus.yml
container_name: prometheus
ports:
- ${PROMETHEUS_PORT}:9091
depends_on:
- openim-server
command: --web.listen-address=:9091 --config.file="/etc/prometheus/prometheus.yml"
networks:
openim-server:
ipv4_address: ${PROMETHEUS_NETWORK_ADDRESS}
grafana:
image: grafana/grafana
volumes:
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/datasource-compose.yaml:/etc/grafana/provisioning/datasources/datasource.yaml
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/grafana.ini:/etc/grafana/grafana.ini
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/node-exporter-full_rev1.json:/var/lib/grafana/dashboards/node-exporter-full_rev1.json
container_name: grafana
ports:
- ${GRAFANA_PORT}:3000
depends_on:
- prometheus
networks:
openim-server:
ipv4_address: ${GRAFANA_NETWORK_ADDRESS}
# node-exporter:
# image: quay.io/prometheus/node-exporter
# container_name: node-exporter
# restart: always
# ports:
# - "9100:9100"
```
|
1.0
|
Optimization and Refactoring of Docker and Docker Compose Network Modes - # Optimization and Refactoring of Docker and Docker Compose Network Modes
**Description:**
#### Background:
Our current Docker and Docker Compose setups use a bridge network mode. However, there's a need to optimize and refactor the network modes for better scalability, maintainability, and performance.
#### Issue:
Based on our existing `docker-compose.yml` file (see attached), there are areas that can be potentially improved in terms of networking:
1. **Use of Bridge Mode:** While bridge mode is a default and often suitable for many applications, there might be scenarios where host or overlay networking modes may offer better performance or more suitable isolation.
2. **Network Addressing:** Hardcoding IP addresses (e.g., `${MYSQL_NETWORK_ADDRESS}`, `${MONGO_NETWORK_ADDRESS}`, etc.) may not be scalable. We need a more dynamic way of handling this.
3. **Dependency Management:** While `depends_on` is used, we might want to look into the use of `wait-for-it` or similar scripts to ensure services are fully up and ready before dependants start.
4. **Consistency:** Ensure that the network configurations are consistent across all services. For instance, all services should be part of the necessary networks, and network mode should be uniform unless there's a valid reason for deviations.
#### Proposed Action:
1. Evaluate the current bridge mode against our application's needs. Consider if host, overlay, or macvlan modes might be more appropriate.
2. Investigate using Docker's native DNS for service discovery rather than hardcoding IP addresses.
3. Implement `wait-for-it` or similar scripts for better management of service dependencies.
4. Ensure all services have consistent network configurations, including necessary settings for IPv4 and IPv6 (if applicable).
5. Update and refactor the `docker-compose.yml` file based on the findings and proposed changes.
## Content
```
#fixme Clone openIM Server project before using docker-compose,project address:https://github.com/OpenIMSDK/Open-IM-Server.git
version: '3'
networks:
openim-server:
driver: bridge
ipam:
driver: default
config:
- subnet: '${DOCKER_BRIDGE_SUBNET}'
gateway: '${DOCKER_BRIDGE_GATEWAY}'
services:
mysql:
image: mysql:5.7
ports:
- "${MYSQL_PORT}:3306"
container_name: mysql
volumes:
- "${DATA_DIR}/components/mysql/data:/var/lib/mysql"
- "/etc/localtime:/etc/localtime"
environment:
MYSQL_ROOT_PASSWORD: "${MYSQL_PASSWORD}"
restart: always
networks:
openim-server:
ipv4_address: ${MYSQL_NETWORK_ADDRESS}
mongodb:
image: mongo:6.0.2
ports:
- "${MONGO_PORT}:27017"
container_name: mongo
command: --wiredTigerCacheSizeGB 1 --auth
volumes:
- "${DATA_DIR}/components/mongodb/data/db:/data/db"
- "${DATA_DIR}/components/mongodb/data/logs:/data/logs"
- "${DATA_DIR}/components/mongodb/data/conf:/etc/mongo"
- "./openim-server/${SERVER_BRANCH}/scripts/mongo-init.sh:/docker-entrypoint-initdb.d/mongo-init.sh:ro"
environment:
- TZ=Asia/Shanghai
- wiredTigerCacheSizeGB=1
- MONGO_INITDB_ROOT_USERNAME=${MONGO_USERNAME}
- MONGO_INITDB_ROOT_PASSWORD=${MONGO_PASSWORD}
- MONGO_INITDB_DATABASE=${MONGO_DATABASE}
restart: always
networks:
openim-server:
ipv4_address: ${MONGO_NETWORK_ADDRESS}
redis:
image: redis:7.0.0
container_name: redis
ports:
- "${REDIS_PORT}:6379"
volumes:
- "${DATA_DIR}/components/redis/data:/data"
- "${DATA_DIR}/components/redis/config/redis.conf:/usr/local/redis/config/redis.conf"
environment:
TZ: Asia/Shanghai
restart: always
sysctls:
net.core.somaxconn: 1024
command: redis-server --requirepass ${REDIS_PASSWORD} --appendonly yes
networks:
openim-server:
ipv4_address: ${REDIS_NETWORK_ADDRESS}
zookeeper:
image: bitnami/zookeeper:3.8
container_name: zookeeper
ports:
- target: 2181
published: "${ZOOKEEPER_PORT}"
protocol: tcp
volumes:
- "/etc/localtime:/etc/localtime"
environment:
- ALLOW_ANONYMOUS_LOGIN=yes
- TZ="Asia/Shanghai"
restart: always
networks:
openim-server:
ipv4_address: ${ZOOKEEPER_NETWORK_ADDRESS}
kafka:
image: 'bitnami/kafka:latest'
container_name: kafka
restart: always
ports:
- target: 9092
published: "${KAFKA_PORT}"
protocol: tcp
volumes:
- ./openim-server/${SERVER_BRANCH}/scripts/create_topic.sh:/opt/bitnami/kafka/create_topic.sh
# - ${DATA_DIR}/components/kafka/data:/bitnami/kafka/data
command: >
bash -c "
/opt/bitnami/scripts/kafka/run.sh & sleep 30; /opt/bitnami/kafka/create_topic.sh; wait
"
environment:
- KAFKA_CFG_NODE_ID=0
- TZ=Asia/Shanghai
- KAFKA_CFG_PROCESS_ROLES=controller,broker
- KAFKA_CFG_LISTENERS=PLAINTEXT://:9092,CONTROLLER://:9093
- KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP=CONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT
- KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@kafka:9093
- KAFKA_CFG_CONTROLLER_LISTENER_NAMES=CONTROLLER
networks:
openim-server:
ipv4_address: ${KAFKA_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
minio:
image: minio/minio
ports:
- "${MINIO_PORT}:9000"
- "9090:9090"
container_name: minio
volumes:
- "${DATA_DIR}/components/mnt/data:/data"
- "${DATA_DIR}/components/mnt/config:/root/.minio"
environment:
MINIO_ROOT_USER: "${MINIO_ACCESS_KEY}"
MINIO_ROOT_PASSWORD: "${MINIO_SECRET_KEY}"
restart: always
command: minio server /data --console-address ':9090'
networks:
openim-server:
ipv4_address: ${MINIO_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-web:
image: ghcr.io/openimsdk/openim-web:latest
container_name: openim-web
environment:
- OPENIM_WEB_DIST_PATH=${OPENIM_WEB_DIST_PATH}
- OPENIM_WEB_PORT=${OPENIM_WEB_PORT}
restart: always
ports:
- "${OPENIM_WEB_PORT}:11001"
networks:
openim-server:
ipv4_address: ${OPENIM_WEB_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-server:
# image: ghcr.io/openimsdk/openim-server:${SERVER_BRANCH}
image: registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-server:${SERVER_BRANCH}
# image: openim/openim-server:${SERVER_BRANCH}
# build: .
container_name: openim-server
ports:
- ${OPENIM_WS_PORT}:10001
- ${API_OPENIM_PORT}:10002
healthcheck:
test: ["CMD", "/openim/openim-server/scripts/check-all.sh"]
interval: 300s
timeout: 10s
retries: 5
volumes:
- ./openim-server/logs:/openim/openim-server/logs
- ./openim-server/_output:/openim/openim-server/_output
- ./openim-server/${SERVER_BRANCH}/config:/openim/openim-server/config
- ./openim-server/${SERVER_BRANCH}/scripts:/openim/openim-server/scripts
restart: always
depends_on:
- kafka
- mysql
- mongodb
- redis
- minio
logging:
driver: json-file
options:
max-size: "1g"
max-file: "2"
networks:
openim-server:
ipv4_address: ${OPENIM_SERVER_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
openim-chat:
# image: ghcr.io/openimsdk/openim-chat:${CHAT_BRANCH}
image: registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-chat:${CHAT_BRANCH}
# image: ghcr.io/openimsdk/openim-chat:${CHAT_BRANCH}
container_name: openim-chat
healthcheck:
test: ["CMD", "/openim/openim-chat/scripts/check_all.sh"]
interval: 300s
timeout: 10s
retries: 5
ports:
- ${OPENIM_CHAT_API_PORT}:10008
- ${OPENIM_ADMIN_API_PORT}:10009
volumes:
- ./openim-chat/logs:/openim/openim-chat/logs
- ./openim-chat/_output:/openim/openim-chat/_output
- ./openim-chat/${CHAT_BRANCH}/config:/openim/openim-chat/config
- ./openim-chat/${CHAT_BRANCH}/scripts:/openim/openim-chat/scripts
restart: always
user: root:root
depends_on:
- mysql
- mongodb
- redis
- minio
- openim-server
logging:
driver: json-file
options:
max-size: "1g"
max-file: "2"
networks:
openim-server:
ipv4_address: ${OPENIM_CHAT_NETWORK_ADDRESS}
extra_hosts:
- "host.docker.internal:host-gateway"
prometheus:
image: prom/prometheus
volumes:
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/prometheus-compose.yml:/etc/prometheus/prometheus.yml
container_name: prometheus
ports:
- ${PROMETHEUS_PORT}:9091
depends_on:
- openim-server
command: --web.listen-address=:9091 --config.file="/etc/prometheus/prometheus.yml"
networks:
openim-server:
ipv4_address: ${PROMETHEUS_NETWORK_ADDRESS}
grafana:
image: grafana/grafana
volumes:
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/datasource-compose.yaml:/etc/grafana/provisioning/datasources/datasource.yaml
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/grafana.ini:/etc/grafana/grafana.ini
- ./openim-server/${SERVER_BRANCH}/.docker-compose_cfg/node-exporter-full_rev1.json:/var/lib/grafana/dashboards/node-exporter-full_rev1.json
container_name: grafana
ports:
- ${GRAFANA_PORT}:3000
depends_on:
- prometheus
networks:
openim-server:
ipv4_address: ${GRAFANA_NETWORK_ADDRESS}
# node-exporter:
# image: quay.io/prometheus/node-exporter
# container_name: node-exporter
# restart: always
# ports:
# - "9100:9100"
```
|
non_process
|
optimization and refactoring of docker and docker compose network modes optimization and refactoring of docker and docker compose network modes description background our current docker and docker compose setups use a bridge network mode however there s a need to optimize and refactor the network modes for better scalability maintainability and performance issue based on our existing docker compose yml file see attached there are areas that can be potentially improved in terms of networking use of bridge mode while bridge mode is a default and often suitable for many applications there might be scenarios where host or overlay networking modes may offer better performance or more suitable isolation network addressing hardcoding ip addresses e g mysql network address mongo network address etc may not be scalable we need a more dynamic way of handling this dependency management while depends on is used we might want to look into the use of wait for it or similar scripts to ensure services are fully up and ready before dependants start consistency ensure that the network configurations are consistent across all services for instance all services should be part of the necessary networks and network mode should be uniform unless there s a valid reason for deviations proposed action evaluate the current bridge mode against our application s needs consider if host overlay or macvlan modes might be more appropriate investigate using docker s native dns for service discovery rather than hardcoding ip addresses implement wait for it or similar scripts for better management of service dependencies ensure all services have consistent network configurations including necessary settings for and if applicable update and refactor the docker compose yml file based on the findings and proposed changes content fixme clone openim server project before using docker compose project address: version networks openim server driver bridge ipam driver default config subnet docker bridge subnet gateway docker bridge gateway services mysql image mysql ports mysql port container name mysql volumes data dir components mysql data var lib mysql etc localtime etc localtime environment mysql root password mysql password restart always networks openim server address mysql network address mongodb image mongo ports mongo port container name mongo command wiredtigercachesizegb auth volumes data dir components mongodb data db data db data dir components mongodb data logs data logs data dir components mongodb data conf etc mongo openim server server branch scripts mongo init sh docker entrypoint initdb d mongo init sh ro environment tz asia shanghai wiredtigercachesizegb mongo initdb root username mongo username mongo initdb root password mongo password mongo initdb database mongo database restart always networks openim server address mongo network address redis image redis container name redis ports redis port volumes data dir components redis data data data dir components redis config redis conf usr local redis config redis conf environment tz asia shanghai restart always sysctls net core somaxconn command redis server requirepass redis password appendonly yes networks openim server address redis network address zookeeper image bitnami zookeeper container name zookeeper ports target published zookeeper port protocol tcp volumes etc localtime etc localtime environment allow anonymous login yes tz asia shanghai restart always networks openim server address zookeeper network address kafka image bitnami kafka latest container name kafka restart always ports target published kafka port protocol tcp volumes openim server server branch scripts create topic sh opt bitnami kafka create topic sh data dir components kafka data bitnami kafka data command bash c opt bitnami scripts kafka run sh sleep opt bitnami kafka create topic sh wait environment kafka cfg node id tz asia shanghai kafka cfg process roles controller broker kafka cfg listeners plaintext controller kafka cfg listener security protocol map controller plaintext plaintext plaintext kafka cfg controller quorum voters kafka kafka cfg controller listener names controller networks openim server address kafka network address extra hosts host docker internal host gateway minio image minio minio ports minio port container name minio volumes data dir components mnt data data data dir components mnt config root minio environment minio root user minio access key minio root password minio secret key restart always command minio server data console address networks openim server address minio network address extra hosts host docker internal host gateway openim web image ghcr io openimsdk openim web latest container name openim web environment openim web dist path openim web dist path openim web port openim web port restart always ports openim web port networks openim server address openim web network address extra hosts host docker internal host gateway openim server image ghcr io openimsdk openim server server branch image registry cn hangzhou aliyuncs com openimsdk openim server server branch image openim openim server server branch build container name openim server ports openim ws port api openim port healthcheck test interval timeout retries volumes openim server logs openim openim server logs openim server output openim openim server output openim server server branch config openim openim server config openim server server branch scripts openim openim server scripts restart always depends on kafka mysql mongodb redis minio logging driver json file options max size max file networks openim server address openim server network address extra hosts host docker internal host gateway openim chat image ghcr io openimsdk openim chat chat branch image registry cn hangzhou aliyuncs com openimsdk openim chat chat branch image ghcr io openimsdk openim chat chat branch container name openim chat healthcheck test interval timeout retries ports openim chat api port openim admin api port volumes openim chat logs openim openim chat logs openim chat output openim openim chat output openim chat chat branch config openim openim chat config openim chat chat branch scripts openim openim chat scripts restart always user root root depends on mysql mongodb redis minio openim server logging driver json file options max size max file networks openim server address openim chat network address extra hosts host docker internal host gateway prometheus image prom prometheus volumes openim server server branch docker compose cfg prometheus compose yml etc prometheus prometheus yml container name prometheus ports prometheus port depends on openim server command web listen address config file etc prometheus prometheus yml networks openim server address prometheus network address grafana image grafana grafana volumes openim server server branch docker compose cfg datasource compose yaml etc grafana provisioning datasources datasource yaml openim server server branch docker compose cfg grafana ini etc grafana grafana ini openim server server branch docker compose cfg node exporter full json var lib grafana dashboards node exporter full json container name grafana ports grafana port depends on prometheus networks openim server address grafana network address node exporter image quay io prometheus node exporter container name node exporter restart always ports
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.