Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
19,667
26,027,589,122
IssuesEvent
2022-12-21 17:43:25
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
opened
Cnvert the python list to the numpy array for the python embedding at rhe base class.
type: enhancement priority: medium alert: NEED MORE DEFINITION alert: NEED ACCOUNT KEY component: CI/CD alert: NEED PROJECT ASSIGNMENT reporting: DTC NCAR Base requestor: METplus Team MET: PreProcessing Tools (Point)
The python list and the numpy array are accepted for the python embedding. The current implementation gives a INFO message if the python list contains numpy data type members. - no info message with 1) python list & python general numeric data type members and 2) numpy array & numpy data type members - info message with the python list & the numpy data type members ``` ==INFO_PYTHON== Recommend using numpy instead of python list for obs_hgt (<class 'numpy.float32'>) ==INFO_PYTHON== Recommend using numpy instead of python list for obs_val (<class 'numpy.float32'>) ``` ## Describe the Enhancement ## MET codes handle better with the numpy array than the python list for the python embedding. It's better to convert the python list to numpy array instead of giving a INFO message. ### Time Estimate ### *Estimate the amount of work required here.* 4 hours ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Cnvert the python list to the numpy array for the python embedding at rhe base class. - The python list and the numpy array are accepted for the python embedding. The current implementation gives a INFO message if the python list contains numpy data type members. - no info message with 1) python list & python general numeric data type members and 2) numpy array & numpy data type members - info message with the python list & the numpy data type members ``` ==INFO_PYTHON== Recommend using numpy instead of python list for obs_hgt (<class 'numpy.float32'>) ==INFO_PYTHON== Recommend using numpy instead of python list for obs_val (<class 'numpy.float32'>) ``` ## Describe the Enhancement ## MET codes handle better with the numpy array than the python list for the python embedding. It's better to convert the python list to numpy array instead of giving a INFO message. ### Time Estimate ### *Estimate the amount of work required here.* 4 hours ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
process
cnvert the python list to the numpy array for the python embedding at rhe base class the python list and the numpy array are accepted for the python embedding the current implementation gives a info message if the python list contains numpy data type members no info message with python list python general numeric data type members and numpy array numpy data type members info message with the python list the numpy data type members info python recommend using numpy instead of python list for obs hgt info python recommend using numpy instead of python list for obs val describe the enhancement met codes handle better with the numpy array than the python list for the python embedding it s better to convert the python list to numpy array instead of giving a info message time estimate estimate the amount of work required here hours sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and development issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
9,296
12,308,143,720
IssuesEvent
2020-05-12 06:35:16
atlesn/rrr
https://api.github.com/repos/atlesn/rrr
opened
Rewrite python module
bug process
python branch. Python module is defunct with new fork system. Code must be re-organized. - All python3 code must reside within the process/source forks, nothing in main thread - Reduce python3 socket class to a wrapper which puts messages into mmap channel
1.0
Rewrite python module - python branch. Python module is defunct with new fork system. Code must be re-organized. - All python3 code must reside within the process/source forks, nothing in main thread - Reduce python3 socket class to a wrapper which puts messages into mmap channel
process
rewrite python module python branch python module is defunct with new fork system code must be re organized all code must reside within the process source forks nothing in main thread reduce socket class to a wrapper which puts messages into mmap channel
1
7,845
11,015,559,365
IssuesEvent
2019-12-05 02:03:51
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
Add a run test flag, to re-run samples.
testing type: process
Sometimes there is a problem with the resources used rather than the code itself, rather than making up a change to get tests to run again, adding a run flag would be nice.
1.0
Add a run test flag, to re-run samples. - Sometimes there is a problem with the resources used rather than the code itself, rather than making up a change to get tests to run again, adding a run flag would be nice.
process
add a run test flag to re run samples sometimes there is a problem with the resources used rather than the code itself rather than making up a change to get tests to run again adding a run flag would be nice
1
12,324
14,880,872,619
IssuesEvent
2021-01-20 09:46:08
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Adjust the space between the icon and the text
Bug P2 Participant manager Process: Fixed Process: Tested dev
[PM] Adjust the space between the icon and text ![oauth](https://user-images.githubusercontent.com/71445210/104905422-53c92100-59a8-11eb-9af5-53b6db88fe33.png)
2.0
[PM] Adjust the space between the icon and the text - [PM] Adjust the space between the icon and text ![oauth](https://user-images.githubusercontent.com/71445210/104905422-53c92100-59a8-11eb-9af5-53b6db88fe33.png)
process
adjust the space between the icon and the text adjust the space between the icon and text
1
17,803
23,728,933,925
IssuesEvent
2022-08-30 22:46:41
googleapis/gapic-generator-python
https://api.github.com/repos/googleapis/gapic-generator-python
opened
Generator itself should run generated unit tests on generated clients
type: process priority: p2
We run generated unit tests in the individual GAPIC repos, and that is as it should be. However, we should also run generated unit tests upstream in the process, as part of the generator CI. These tests should ideally encompass the full matrix: ``` {regular templates, Ads templates} × {transport=rest, transport=grpc} × (for rest) {numericEnums on, numericEnums off} ``` though perhaps some combinations are easily doable at the moment and may need to be deferred. The most obvious direction here is to do this using the Showcase protos as much as possible. (This issue created as a refinement of #1404)
1.0
Generator itself should run generated unit tests on generated clients - We run generated unit tests in the individual GAPIC repos, and that is as it should be. However, we should also run generated unit tests upstream in the process, as part of the generator CI. These tests should ideally encompass the full matrix: ``` {regular templates, Ads templates} × {transport=rest, transport=grpc} × (for rest) {numericEnums on, numericEnums off} ``` though perhaps some combinations are easily doable at the moment and may need to be deferred. The most obvious direction here is to do this using the Showcase protos as much as possible. (This issue created as a refinement of #1404)
process
generator itself should run generated unit tests on generated clients we run generated unit tests in the individual gapic repos and that is as it should be however we should also run generated unit tests upstream in the process as part of the generator ci these tests should ideally encompass the full matrix regular templates ads templates × transport rest transport grpc × for rest numericenums on numericenums off though perhaps some combinations are easily doable at the moment and may need to be deferred the most obvious direction here is to do this using the showcase protos as much as possible this issue created as a refinement of
1
134,490
5,227,347,124
IssuesEvent
2017-01-28 01:17:10
mgordon34/Conversation-Analysis
https://api.github.com/repos/mgordon34/Conversation-Analysis
reopened
As a user, I want to know the inputs and outputs of the algorithm. #3a
Back End Feature High Priority
## Story/Task Details - [x] Obtained parsed text - [x] Display an output from parsed text ## Acceptance Scenarios - Given: Parsed text - When: User selects to analyze - Then: Display information based on text ## Done Done Criteria Output is displayed on analysis page
1.0
As a user, I want to know the inputs and outputs of the algorithm. #3a - ## Story/Task Details - [x] Obtained parsed text - [x] Display an output from parsed text ## Acceptance Scenarios - Given: Parsed text - When: User selects to analyze - Then: Display information based on text ## Done Done Criteria Output is displayed on analysis page
non_process
as a user i want to know the inputs and outputs of the algorithm story task details obtained parsed text display an output from parsed text acceptance scenarios given parsed text when user selects to analyze then display information based on text done done criteria output is displayed on analysis page
0
17,358
23,185,320,704
IssuesEvent
2022-08-01 07:52:55
LAAC-LSCP/ChildProject
https://api.github.com/repos/LAAC-LSCP/ChildProject
closed
drop --split ?
question audio-processing
**Is your feature request related to a problem? Please describe.** `--split` is: - not fully supported yet because you cannot use split recordings in ChildProject's routines (it assumes a 1:1 mapping between raw and converted files) - apparently not as critical since ELAN seems to support longer audio recordings now ? Is this feature still really useful ? If not, I'd vote for dropping it
1.0
drop --split ? - **Is your feature request related to a problem? Please describe.** `--split` is: - not fully supported yet because you cannot use split recordings in ChildProject's routines (it assumes a 1:1 mapping between raw and converted files) - apparently not as critical since ELAN seems to support longer audio recordings now ? Is this feature still really useful ? If not, I'd vote for dropping it
process
drop split is your feature request related to a problem please describe split is not fully supported yet because you cannot use split recordings in childproject s routines it assumes a mapping between raw and converted files apparently not as critical since elan seems to support longer audio recordings now is this feature still really useful if not i d vote for dropping it
1
16,897
22,197,231,525
IssuesEvent
2022-06-07 08:05:07
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
BrokerReprocessingTest.shouldCorrelateMessageAfterRestartIfPublishedBefore is flaky
kind/flake team/distributed team/process-automation area/test
**Summary** - How often does the test fail? - at least 5 times so far - Does it block your work? - no, but it's a little annoying - Do we suspect that it is a real failure? - no, it's either just flaky or something with the new pipeline **Failures** > Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins <details><summary>Example assertion failure</summary> <pre> io.camunda.zeebe.client.api.command.ClientStatusException: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.transformExecutionException(ZeebeClientFutureImpl.java:93) at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.join(ZeebeClientFutureImpl.java:50) at io.camunda.zeebe.it.startup.BrokerReprocessingTest.startProcessInstance(BrokerReprocessingTest.java:593) at io.camunda.zeebe.it.startup.BrokerReprocessingTest.shouldCorrelateMessageAfterRestartIfPublishedBefore(BrokerReprocessingTest.java:539) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:258) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.ExecutionException: io.grpc.StatusRuntimeException: INTERNAL: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2073) at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.join(ZeebeClientFutureImpl.java:48) ... 19 more Caused by: io.grpc.StatusRuntimeException: INTERNAL: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at io.grpc.Status.asRuntimeException(Status.java:535) at io.grpc.stub.ClientCalls$StreamObserverToCallListenerAdapter.onClose(ClientCalls.java:487) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ... 1 more </pre> </details>
1.0
BrokerReprocessingTest.shouldCorrelateMessageAfterRestartIfPublishedBefore is flaky - **Summary** - How often does the test fail? - at least 5 times so far - Does it block your work? - no, but it's a little annoying - Do we suspect that it is a real failure? - no, it's either just flaky or something with the new pipeline **Failures** > Outline known failure cases, e.g. a failed assertion and its stacktrace obtained from Jenkins <details><summary>Example assertion failure</summary> <pre> io.camunda.zeebe.client.api.command.ClientStatusException: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.transformExecutionException(ZeebeClientFutureImpl.java:93) at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.join(ZeebeClientFutureImpl.java:50) at io.camunda.zeebe.it.startup.BrokerReprocessingTest.startProcessInstance(BrokerReprocessingTest.java:593) at io.camunda.zeebe.it.startup.BrokerReprocessingTest.shouldCorrelateMessageAfterRestartIfPublishedBefore(BrokerReprocessingTest.java:539) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:258) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.ExecutionException: io.grpc.StatusRuntimeException: INTERNAL: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2073) at io.camunda.zeebe.client.impl.ZeebeClientFutureImpl.join(ZeebeClientFutureImpl.java:48) ... 19 more Caused by: io.grpc.StatusRuntimeException: INTERNAL: Unexpected error occurred during the request processing: No remote message handler registered for this message, subject command-api-1 at io.grpc.Status.asRuntimeException(Status.java:535) at io.grpc.stub.ClientCalls$StreamObserverToCallListenerAdapter.onClose(ClientCalls.java:487) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ... 1 more </pre> </details>
process
brokerreprocessingtest shouldcorrelatemessageafterrestartifpublishedbefore is flaky summary how often does the test fail at least times so far does it block your work no but it s a little annoying do we suspect that it is a real failure no it s either just flaky or something with the new pipeline failures outline known failure cases e g a failed assertion and its stacktrace obtained from jenkins example assertion failure io camunda zeebe client api command clientstatusexception unexpected error occurred during the request processing no remote message handler registered for this message subject command api at io camunda zeebe client impl zeebeclientfutureimpl transformexecutionexception zeebeclientfutureimpl java at io camunda zeebe client impl zeebeclientfutureimpl join zeebeclientfutureimpl java at io camunda zeebe it startup brokerreprocessingtest startprocessinstance brokerreprocessingtest java at io camunda zeebe it startup brokerreprocessingtest shouldcorrelatemessageafterrestartifpublishedbefore brokerreprocessingtest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit rules expectedexception expectedexceptionstatement evaluate expectedexception java at org junit rules externalresource evaluate externalresource java at org junit rules testwatcher evaluate testwatcher java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java base java util concurrent futuretask run futuretask java at java base java lang thread run thread java caused by java util concurrent executionexception io grpc statusruntimeexception internal unexpected error occurred during the request processing no remote message handler registered for this message subject command api at java base java util concurrent completablefuture reportget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io camunda zeebe client impl zeebeclientfutureimpl join zeebeclientfutureimpl java more caused by io grpc statusruntimeexception internal unexpected error occurred during the request processing no remote message handler registered for this message subject command api at io grpc status asruntimeexception status java at io grpc stub clientcalls streamobservertocalllisteneradapter onclose clientcalls java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java more
1
413
2,852,061,348
IssuesEvent
2015-06-01 11:18:04
genomizer/genomizer-server
https://api.github.com/repos/genomizer/genomizer-server
closed
Vagrant: /tmp and /data are on different partitions
Data Storage enhancement High priority Processing
``` $ mount [...] /dev/sdb2 on / type ext4 (rw,relatime,seclabel,data=ordered) [...] /dev/sda1 on /Data type ext4 (rw,relatime,seclabel,data=ordered) [...] ``` This means that uploaded files have to be copied, which takes a long time if the file in question is large. If /tmp was on the same partition as /data, they could be just renamed instead.
1.0
Vagrant: /tmp and /data are on different partitions - ``` $ mount [...] /dev/sdb2 on / type ext4 (rw,relatime,seclabel,data=ordered) [...] /dev/sda1 on /Data type ext4 (rw,relatime,seclabel,data=ordered) [...] ``` This means that uploaded files have to be copied, which takes a long time if the file in question is large. If /tmp was on the same partition as /data, they could be just renamed instead.
process
vagrant tmp and data are on different partitions mount dev on type rw relatime seclabel data ordered dev on data type rw relatime seclabel data ordered this means that uploaded files have to be copied which takes a long time if the file in question is large if tmp was on the same partition as data they could be just renamed instead
1
605,353
18,734,122,967
IssuesEvent
2021-11-04 03:42:54
squaresLab/security-repair-benchmarks
https://api.github.com/repos/squaresLab/security-repair-benchmarks
closed
Incorrect trace for CVE-2012-5134
bug high-priority
The fault location is in the method `xmlParseAttValueComplex` [https://gitlab.gnome.org/GNOME/libxml2/-/blob/6a36fbe3b3e001a8a840b5c1fdd81cefc9947f0d/parser.c#L3894] which is not in the trace in the trace file. I guess the same problem is also with other XML subjects. Trace file for the function [xmllint.bbt.txt](https://github.com/squaresLab/security-repair-benchmarks/files/7365166/xmllint.bbt.txt). As the tracer simply logs the executed basic blocks, I believe the problem is with the configuration to run the tracer (wllvm, etc.) or with the exploit itself. `input.json` has no role in this issue. The file is not used by tracer. @ChrisTimperley Let me know if you have time to debug it. We can schedule a debugging session if you are interested. @mig40000 @yannicnoller FYI.
1.0
Incorrect trace for CVE-2012-5134 - The fault location is in the method `xmlParseAttValueComplex` [https://gitlab.gnome.org/GNOME/libxml2/-/blob/6a36fbe3b3e001a8a840b5c1fdd81cefc9947f0d/parser.c#L3894] which is not in the trace in the trace file. I guess the same problem is also with other XML subjects. Trace file for the function [xmllint.bbt.txt](https://github.com/squaresLab/security-repair-benchmarks/files/7365166/xmllint.bbt.txt). As the tracer simply logs the executed basic blocks, I believe the problem is with the configuration to run the tracer (wllvm, etc.) or with the exploit itself. `input.json` has no role in this issue. The file is not used by tracer. @ChrisTimperley Let me know if you have time to debug it. We can schedule a debugging session if you are interested. @mig40000 @yannicnoller FYI.
non_process
incorrect trace for cve the fault location is in the method xmlparseattvaluecomplex which is not in the trace in the trace file i guess the same problem is also with other xml subjects trace file for the function as the tracer simply logs the executed basic blocks i believe the problem is with the configuration to run the tracer wllvm etc or with the exploit itself input json has no role in this issue the file is not used by tracer christimperley let me know if you have time to debug it we can schedule a debugging session if you are interested yannicnoller fyi
0
402,896
27,392,668,631
IssuesEvent
2023-02-28 17:17:14
mikakaraila/node-red-contrib-opcua
https://api.github.com/repos/mikakaraila/node-red-contrib-opcua
closed
Can't create polling model with READ option
documentation
Hi, I want to create a flow for polling data every 5 seconds from 93 nodes. The subscription model doesn't work for me because I always end up with "BadTooManySessions" and also there are some metrics that change very selden, so I don't see anything in my Grafana Dashboard. I also noticed that if I loop the inject node I always also start the Node-Browser again and I also end up with "BadTooManySessions". Whats the best way to create a simple polling flow? Thanks in advance!
1.0
Can't create polling model with READ option - Hi, I want to create a flow for polling data every 5 seconds from 93 nodes. The subscription model doesn't work for me because I always end up with "BadTooManySessions" and also there are some metrics that change very selden, so I don't see anything in my Grafana Dashboard. I also noticed that if I loop the inject node I always also start the Node-Browser again and I also end up with "BadTooManySessions". Whats the best way to create a simple polling flow? Thanks in advance!
non_process
can t create polling model with read option hi i want to create a flow for polling data every seconds from nodes the subscription model doesn t work for me because i always end up with badtoomanysessions and also there are some metrics that change very selden so i don t see anything in my grafana dashboard i also noticed that if i loop the inject node i always also start the node browser again and i also end up with badtoomanysessions whats the best way to create a simple polling flow thanks in advance
0
4,831
7,725,942,661
IssuesEvent
2018-05-24 19:35:51
kaching-hq/Privacy-and-Security
https://api.github.com/repos/kaching-hq/Privacy-and-Security
closed
Lay the groundwork for the DPO role
Processes
- [x] Designate a DPO - [x] Communicate decision to Datainspektionen - [x] Document decision (auditing and board purposes) - [x] Choose a DPMS
1.0
Lay the groundwork for the DPO role - - [x] Designate a DPO - [x] Communicate decision to Datainspektionen - [x] Document decision (auditing and board purposes) - [x] Choose a DPMS
process
lay the groundwork for the dpo role designate a dpo communicate decision to datainspektionen document decision auditing and board purposes choose a dpms
1
8,766
11,884,486,907
IssuesEvent
2020-03-27 17:44:30
fraction/oasis
https://api.github.com/repos/fraction/oasis
closed
Chatting about Oasis
process
## What's the problem you want solved? There's no place to chat about Oasis. There are GitHub issues for tracking problems and SSB messages, but neither feel very lightweight for casual conversation. ## Is there a solution you'd like to recommend? I've started a Matrix room so that people can stop by and chat: https://riot.im/app/#/room/!YQpqIZlvBVPYRwrkXp:matrix.org Maybe we could/should add it to the docs? I was originally thinking that a Signal group would be nice, but it's not as easy to join and requires phone numbers. Is Matrix is a good compromise? ## What version or commit of Oasis are you using? latest
1.0
Chatting about Oasis - ## What's the problem you want solved? There's no place to chat about Oasis. There are GitHub issues for tracking problems and SSB messages, but neither feel very lightweight for casual conversation. ## Is there a solution you'd like to recommend? I've started a Matrix room so that people can stop by and chat: https://riot.im/app/#/room/!YQpqIZlvBVPYRwrkXp:matrix.org Maybe we could/should add it to the docs? I was originally thinking that a Signal group would be nice, but it's not as easy to join and requires phone numbers. Is Matrix is a good compromise? ## What version or commit of Oasis are you using? latest
process
chatting about oasis what s the problem you want solved there s no place to chat about oasis there are github issues for tracking problems and ssb messages but neither feel very lightweight for casual conversation is there a solution you d like to recommend i ve started a matrix room so that people can stop by and chat maybe we could should add it to the docs i was originally thinking that a signal group would be nice but it s not as easy to join and requires phone numbers is matrix is a good compromise what version or commit of oasis are you using latest
1
8,432
3,739,175,101
IssuesEvent
2016-03-09 02:45:47
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
ReadyToRun: NullReferenceException in superlong tests
bug CodeGen ReadyToRun
Tests: JIT\Methodical\int64\superlong\_dbgsuperlong\_dbgsuperlong.cmd JIT\Methodical\int64\superlong\_relsuperlong\_relsuperlong.cmd JIT\Methodical\int64\superlong\_il_relsuperlong\_il_relsuperlong.cmd JIT\Methodical\int64\superlong\_il_dbgsuperlong\_il_dbgsuperlong.cmd Repro steps: Git commit hash: c89bf3f E:\coreclr>rd /s/q E:\coreclr\bin\tests E:\coreclr>build checked skiptests E:\coreclr>tests\buildtest.cmd checked crossgen E:\coreclr>tests\runtest.cmd checked Unhandled Exception: System.NullReferenceException: Object reference not set to an instance of an object. at JitTest.superlong.add(superlong op1, superlong op2) at JitTest.superlong.Main()
1.0
ReadyToRun: NullReferenceException in superlong tests - Tests: JIT\Methodical\int64\superlong\_dbgsuperlong\_dbgsuperlong.cmd JIT\Methodical\int64\superlong\_relsuperlong\_relsuperlong.cmd JIT\Methodical\int64\superlong\_il_relsuperlong\_il_relsuperlong.cmd JIT\Methodical\int64\superlong\_il_dbgsuperlong\_il_dbgsuperlong.cmd Repro steps: Git commit hash: c89bf3f E:\coreclr>rd /s/q E:\coreclr\bin\tests E:\coreclr>build checked skiptests E:\coreclr>tests\buildtest.cmd checked crossgen E:\coreclr>tests\runtest.cmd checked Unhandled Exception: System.NullReferenceException: Object reference not set to an instance of an object. at JitTest.superlong.add(superlong op1, superlong op2) at JitTest.superlong.Main()
non_process
readytorun nullreferenceexception in superlong tests tests jit methodical superlong dbgsuperlong dbgsuperlong cmd jit methodical superlong relsuperlong relsuperlong cmd jit methodical superlong il relsuperlong il relsuperlong cmd jit methodical superlong il dbgsuperlong il dbgsuperlong cmd repro steps git commit hash e coreclr rd s q e coreclr bin tests e coreclr build checked skiptests e coreclr tests buildtest cmd checked crossgen e coreclr tests runtest cmd checked unhandled exception system nullreferenceexception object reference not set to an instance of an object at jittest superlong add superlong superlong at jittest superlong main
0
15,872
3,343,922,578
IssuesEvent
2015-11-15 21:44:51
logista/btsite2015
https://api.github.com/repos/logista/btsite2015
closed
footnotes need to be tweaked
design
They seem a bit big, too close together, and I'd like a separator of some sort.
1.0
footnotes need to be tweaked - They seem a bit big, too close together, and I'd like a separator of some sort.
non_process
footnotes need to be tweaked they seem a bit big too close together and i d like a separator of some sort
0
194,906
6,900,465,716
IssuesEvent
2017-11-24 18:56:13
levex/group-project
https://api.github.com/repos/levex/group-project
opened
Refactor frontend
category: cleanup category: improvement component: client difficulty: easy priority: medium requires: work
Frontend is a mess, needs clean up - break it down into JSX components and JS app logic.
1.0
Refactor frontend - Frontend is a mess, needs clean up - break it down into JSX components and JS app logic.
non_process
refactor frontend frontend is a mess needs clean up break it down into jsx components and js app logic
0
9,843
12,835,870,002
IssuesEvent
2020-07-07 13:33:11
keep-network/keep-core
https://api.github.com/repos/keep-network/keep-core
closed
Make sure button hover states are consistent to Design System figma file
:old_key: token dashboard process & client team
There are a few instances of button states that deviate from Figma in the dApp. An example screenshot below: <img width="442" alt="KEEP_Token_Dashboard" src="https://user-images.githubusercontent.com/57226633/81453449-90752500-9157-11ea-94cb-9a69fa1117b2.png"> Here's how the button states should be: <img width="867" alt="Keep_Design_System" src="https://user-images.githubusercontent.com/57226633/81453675-36c12a80-9158-11ea-870b-5a73d2adc85e.png"> Button states can be referred to here in the design system in Figma (hoping to get this into a more dev friendly format at some point, but for now it's a component based Figma file) https://www.figma.com/file/ARAMFLMOPal0JewjvlZGhX/Keep-Design-System?node-id=787%3A1141
1.0
Make sure button hover states are consistent to Design System figma file - There are a few instances of button states that deviate from Figma in the dApp. An example screenshot below: <img width="442" alt="KEEP_Token_Dashboard" src="https://user-images.githubusercontent.com/57226633/81453449-90752500-9157-11ea-94cb-9a69fa1117b2.png"> Here's how the button states should be: <img width="867" alt="Keep_Design_System" src="https://user-images.githubusercontent.com/57226633/81453675-36c12a80-9158-11ea-870b-5a73d2adc85e.png"> Button states can be referred to here in the design system in Figma (hoping to get this into a more dev friendly format at some point, but for now it's a component based Figma file) https://www.figma.com/file/ARAMFLMOPal0JewjvlZGhX/Keep-Design-System?node-id=787%3A1141
process
make sure button hover states are consistent to design system figma file there are a few instances of button states that deviate from figma in the dapp an example screenshot below img width alt keep token dashboard src here s how the button states should be img width alt keep design system src button states can be referred to here in the design system in figma hoping to get this into a more dev friendly format at some point but for now it s a component based figma file
1
20,265
26,887,736,368
IssuesEvent
2023-02-06 05:45:55
openxla/stablehlo
https://api.github.com/repos/openxla/stablehlo
opened
Share the slide deck used to create images for the spec
Process
After #735 is fixed, we should publish the deck that we used to create images for the spec.
1.0
Share the slide deck used to create images for the spec - After #735 is fixed, we should publish the deck that we used to create images for the spec.
process
share the slide deck used to create images for the spec after is fixed we should publish the deck that we used to create images for the spec
1
250,869
7,988,770,916
IssuesEvent
2018-07-19 12:15:50
MARKETProtocol/dApp
https://api.github.com/repos/MARKETProtocol/dApp
closed
Create a store of available contracts for improved loading speeds
Priority: High Status: Pending Type: Enhancement
<!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### Why Is this Needed? [comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen) *Summary*: Currently when the number of contracts gets large, it begins to take a very long time to build up all the contract data from the blockchain. We want to devise a way to track these contracts in a data store to make loading them more efficient. ### Solution [comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.) *Summary*: Store contracts in a database for easier access. ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) - [ ] Schema - [ ] API - [ ] Create - [ ] Read - [ ] Update
1.0
Create a store of available contracts for improved loading speeds - <!-- Hello! Please use the template below for issue ideas or bugs found within MARKET Protocol. If it is general support you need, reach out to us at https://marketprotocol.io/discord Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> <!-- ## Before you `start work` Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties) If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible. Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline. Please also note that in order for work to be accepted, all code must be accompanied by test cases as well. --> ### Why Is this Needed? [comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen) *Summary*: Currently when the number of contracts gets large, it begins to take a very long time to build up all the contract data from the blockchain. We want to devise a way to track these contracts in a data store to make loading them more efficient. ### Solution [comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.) *Summary*: Store contracts in a database for easier access. ### Definition of Done [comment]: # (Any other information that would be useful, bullets are helpful.) - [ ] Schema - [ ] API - [ ] Create - [ ] Read - [ ] Update
non_process
create a store of available contracts for improved loading speeds hello please use the template below for issue ideas or bugs found within market protocol if it is general support you need reach out to us at provide a general summary of the issue in the title above and use relevant fields below to define the problem before you start work please read our contribution and if there is a bounty involved please also see if you have ongoing work from other bounties with us where funding has not been released please do not pick up a new issue we would like to involve as many contributors as possible and parallelize the work flow as much as possible please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline please also note that in order for work to be accepted all code must be accompanied by test cases as well why is this needed describe the problem and why this task is needed provide description of the current state what you would like to happen and what actually happen summary currently when the number of contracts gets large it begins to take a very long time to build up all the contract data from the blockchain we want to devise a way to track these contracts in a data store to make loading them more efficient solution provide a summary of the solution and a task list on what needs to be fixed summary store contracts in a database for easier access definition of done any other information that would be useful bullets are helpful schema api create read update
0
21,245
28,369,914,355
IssuesEvent
2023-04-12 16:12:15
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Allow additional testkube testsuite executionRequest variables
enhancement test process
### Problem For different environments, we want to set specific variables in testkube testsuite's `.spec.executionRequest.variables` ### Solution Support `.Values.testkube.test.extraExeucitionRequestVariables` config in `hedera-mirror-common` chart, and customize it for different envs if necessary ### Alternatives _No response_
1.0
Allow additional testkube testsuite executionRequest variables - ### Problem For different environments, we want to set specific variables in testkube testsuite's `.spec.executionRequest.variables` ### Solution Support `.Values.testkube.test.extraExeucitionRequestVariables` config in `hedera-mirror-common` chart, and customize it for different envs if necessary ### Alternatives _No response_
process
allow additional testkube testsuite executionrequest variables problem for different environments we want to set specific variables in testkube testsuite s spec executionrequest variables solution support values testkube test extraexeucitionrequestvariables config in hedera mirror common chart and customize it for different envs if necessary alternatives no response
1
8,597
11,758,897,881
IssuesEvent
2020-03-13 16:14:12
nltk/nltk
https://api.github.com/repos/nltk/nltk
closed
Thread-safe CorpusReader
corpus inactive multithread / multiprocessing
I'm using a custom subclassed `CorpusReader` to feed documents to the Word2Vec model in gensim. Because the reader isn't threadsafe, it breaks. I can write my own simple reader for my files, but it would be nice if I could just use nltk. Based on the comments above [iterate_from](https://github.com/nltk/nltk/blob/develop/nltk/corpus/reader/util.py#L259) it looks like some locking could be added to make the reader thread-safe. What exactly what I need to acquire a lock on to make `iterate_from` safe? -- Eric
1.0
Thread-safe CorpusReader - I'm using a custom subclassed `CorpusReader` to feed documents to the Word2Vec model in gensim. Because the reader isn't threadsafe, it breaks. I can write my own simple reader for my files, but it would be nice if I could just use nltk. Based on the comments above [iterate_from](https://github.com/nltk/nltk/blob/develop/nltk/corpus/reader/util.py#L259) it looks like some locking could be added to make the reader thread-safe. What exactly what I need to acquire a lock on to make `iterate_from` safe? -- Eric
process
thread safe corpusreader i m using a custom subclassed corpusreader to feed documents to the model in gensim because the reader isn t threadsafe it breaks i can write my own simple reader for my files but it would be nice if i could just use nltk based on the comments above it looks like some locking could be added to make the reader thread safe what exactly what i need to acquire a lock on to make iterate from safe eric
1
5,612
8,468,916,906
IssuesEvent
2018-10-23 21:07:40
carloseduardov8/Viajato
https://api.github.com/repos/carloseduardov8/Viajato
reopened
Popular base de dados
Epic Priority:Normal Process: Setup Environment
Preencher valores fictícios de linhas aéreas, vôos, locadoras de veículos etc.
1.0
Popular base de dados - Preencher valores fictícios de linhas aéreas, vôos, locadoras de veículos etc.
process
popular base de dados preencher valores fictícios de linhas aéreas vôos locadoras de veículos etc
1
239,372
26,223,404,373
IssuesEvent
2023-01-04 16:31:59
NS-Mend/Java-Demo
https://api.github.com/repos/NS-Mend/Java-Demo
closed
CVE-2021-35043 (Medium) detected in antisamy-1.5.3.jar - autoclosed
security vulnerability
## CVE-2021-35043 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>antisamy-1.5.3.jar</b></p></summary> <p>The OWASP AntiSamy project is a collection of APIs for safely allowing users to supply their own HTML and CSS without exposing the site to XSS vulnerabilities.</p> <p>Library home page: <a href="http://www.owasp.org/index.php/Category:OWASP_AntiSamy_Project">http://www.owasp.org/index.php/Category:OWASP_AntiSamy_Project</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/owasp/antisamy/antisamy/1.5.3/antisamy-1.5.3.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **antisamy-1.5.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/NS-Mend/Java-Demo/commit/7029f3960bcddacd18c3a708c2d968d98d8a978f">7029f3960bcddacd18c3a708c2d968d98d8a978f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OWASP AntiSamy before 1.6.4 allows XSS via HTML attributes when using the HTML output serializer (XHTML is not affected). This was demonstrated by a javascript: URL with &#00058 as the replacement for the : character. <p>Publish Date: 2021-07-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35043>CVE-2021-35043</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-35043">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-35043</a></p> <p>Release Date: 2021-07-19</p> <p>Fix Resolution (org.owasp.antisamy:antisamy): 1.6.4</p> <p>Direct dependency fix Resolution (org.owasp.esapi:esapi): 2.3.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-35043 (Medium) detected in antisamy-1.5.3.jar - autoclosed - ## CVE-2021-35043 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>antisamy-1.5.3.jar</b></p></summary> <p>The OWASP AntiSamy project is a collection of APIs for safely allowing users to supply their own HTML and CSS without exposing the site to XSS vulnerabilities.</p> <p>Library home page: <a href="http://www.owasp.org/index.php/Category:OWASP_AntiSamy_Project">http://www.owasp.org/index.php/Category:OWASP_AntiSamy_Project</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/owasp/antisamy/antisamy/1.5.3/antisamy-1.5.3.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **antisamy-1.5.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/NS-Mend/Java-Demo/commit/7029f3960bcddacd18c3a708c2d968d98d8a978f">7029f3960bcddacd18c3a708c2d968d98d8a978f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OWASP AntiSamy before 1.6.4 allows XSS via HTML attributes when using the HTML output serializer (XHTML is not affected). This was demonstrated by a javascript: URL with &#00058 as the replacement for the : character. <p>Publish Date: 2021-07-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35043>CVE-2021-35043</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-35043">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-35043</a></p> <p>Release Date: 2021-07-19</p> <p>Fix Resolution (org.owasp.antisamy:antisamy): 1.6.4</p> <p>Direct dependency fix Resolution (org.owasp.esapi:esapi): 2.3.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_process
cve medium detected in antisamy jar autoclosed cve medium severity vulnerability vulnerable library antisamy jar the owasp antisamy project is a collection of apis for safely allowing users to supply their own html and css without exposing the site to xss vulnerabilities library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org owasp antisamy antisamy antisamy jar dependency hierarchy esapi jar root library x antisamy jar vulnerable library found in head commit a href found in base branch master vulnerability details owasp antisamy before allows xss via html attributes when using the html output serializer xhtml is not affected this was demonstrated by a javascript url with as the replacement for the character publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org owasp antisamy antisamy direct dependency fix resolution org owasp esapi esapi check this box to open an automated fix pr
0
181,541
30,700,410,631
IssuesEvent
2023-07-26 22:35:50
MozillaFoundation/foundation.mozilla.org
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
closed
[PNI] Add Related Articles to Article Pages
design buyer's guide 🛍
Now that Related Articles have been mostly fixed, we should add some to production. - [x] Add link to previous article input spreadsheet **Goals** - [ ] Add at least three related articles to each article, track progress in this [spreadsheet](https://docs.google.com/spreadsheets/d/1uHfnTP7cyU85qfh8AY8Nk7kuHIcPYjNBnw0aSpfLpvs/edit?usp=sharing).
1.0
[PNI] Add Related Articles to Article Pages - Now that Related Articles have been mostly fixed, we should add some to production. - [x] Add link to previous article input spreadsheet **Goals** - [ ] Add at least three related articles to each article, track progress in this [spreadsheet](https://docs.google.com/spreadsheets/d/1uHfnTP7cyU85qfh8AY8Nk7kuHIcPYjNBnw0aSpfLpvs/edit?usp=sharing).
non_process
add related articles to article pages now that related articles have been mostly fixed we should add some to production add link to previous article input spreadsheet goals add at least three related articles to each article track progress in this
0
315
2,794,858,274
IssuesEvent
2015-05-11 18:55:40
rancherio/rancher
https://api.github.com/repos/rancherio/rancher
opened
Service won't fully start
area/service area/ui bug status/blocker
v0.20.0 Steps: 1. Create a service, scale 1, image hello-world Results: Service won't fully start. Network Agent created, container created but the service won't finish loading. Expected: Service should finish.
1.0
Service won't fully start - v0.20.0 Steps: 1. Create a service, scale 1, image hello-world Results: Service won't fully start. Network Agent created, container created but the service won't finish loading. Expected: Service should finish.
non_process
service won t fully start steps create a service scale image hello world results service won t fully start network agent created container created but the service won t finish loading expected service should finish
0
8,665
11,799,897,723
IssuesEvent
2020-03-18 16:37:25
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
closed
Change default error formatting from `pretty` to `colorless`
kind/improvement process/candidate topic: dx
## Problem When an error is thrown in Prisma Client currently the error is "pretty printed" by default (see [spec](https://github.com/prisma/specs/blob/master/prisma-client-js/README.md#error-formatting)) which is making use of ASCII colors. While this provides a better DX in environments which support ASCII colors, in environments that don't provide support this results in unreadable error messages. See the list of issues below for reference: - https://github.com/prisma/prisma-client-js/issues/187 - https://github.com/prisma/prisma-client-js/issues/367 - https://github.com/prisma/prisma-client-js/issues/491 - https://github.com/prisma/prisma-client-js/issues/235 - https://github.com/prisma/prisma2/issues/1856 - https://github.com/prisma/prisma2/issues/1260 - https://github.com/prisma/prisma2/issues/371 - https://github.com/prisma/prisma2/issues/771 When a user runs into this situation (i.e. an error is thrown in an environment that doesn't support ASCII colors) the error message is barely readable. While it's theoretically possible that in these cases a user could manually disable colored error formatting by setting `errorFormat: 'colorless'` it's an unrealistic assumption that a user would know (or intuitively assume) that the error formatting can be configured in order to get a more readable output. To make matters worse often it might be really hard to reproduce an error due to the non-idempotent nature of most applications so even if a user has figured out how to configure the error formatting correctly it might be a lot of work to reproduce the error. ## Solution I suggest to change the default error formatting from `pretty` to `colorless` in order to avoid the problem described above. I'm all in favor of trying to provide the best possible DX out of the box without further configuration but we need to do so in a **safe** way that avoids negative surprise. Once we've changed the default error formatting, I suggest to collaborate with framework authors (e.g. Redwood, Nexus, ...) so these frameworks enable pretty error formatting out of the box as they can guarantee a safe environment. We should also investigate whether there are better heuristics to safely determine whether an environment support ASCII colors.
1.0
Change default error formatting from `pretty` to `colorless` - ## Problem When an error is thrown in Prisma Client currently the error is "pretty printed" by default (see [spec](https://github.com/prisma/specs/blob/master/prisma-client-js/README.md#error-formatting)) which is making use of ASCII colors. While this provides a better DX in environments which support ASCII colors, in environments that don't provide support this results in unreadable error messages. See the list of issues below for reference: - https://github.com/prisma/prisma-client-js/issues/187 - https://github.com/prisma/prisma-client-js/issues/367 - https://github.com/prisma/prisma-client-js/issues/491 - https://github.com/prisma/prisma-client-js/issues/235 - https://github.com/prisma/prisma2/issues/1856 - https://github.com/prisma/prisma2/issues/1260 - https://github.com/prisma/prisma2/issues/371 - https://github.com/prisma/prisma2/issues/771 When a user runs into this situation (i.e. an error is thrown in an environment that doesn't support ASCII colors) the error message is barely readable. While it's theoretically possible that in these cases a user could manually disable colored error formatting by setting `errorFormat: 'colorless'` it's an unrealistic assumption that a user would know (or intuitively assume) that the error formatting can be configured in order to get a more readable output. To make matters worse often it might be really hard to reproduce an error due to the non-idempotent nature of most applications so even if a user has figured out how to configure the error formatting correctly it might be a lot of work to reproduce the error. ## Solution I suggest to change the default error formatting from `pretty` to `colorless` in order to avoid the problem described above. I'm all in favor of trying to provide the best possible DX out of the box without further configuration but we need to do so in a **safe** way that avoids negative surprise. Once we've changed the default error formatting, I suggest to collaborate with framework authors (e.g. Redwood, Nexus, ...) so these frameworks enable pretty error formatting out of the box as they can guarantee a safe environment. We should also investigate whether there are better heuristics to safely determine whether an environment support ASCII colors.
process
change default error formatting from pretty to colorless problem when an error is thrown in prisma client currently the error is pretty printed by default see which is making use of ascii colors while this provides a better dx in environments which support ascii colors in environments that don t provide support this results in unreadable error messages see the list of issues below for reference when a user runs into this situation i e an error is thrown in an environment that doesn t support ascii colors the error message is barely readable while it s theoretically possible that in these cases a user could manually disable colored error formatting by setting errorformat colorless it s an unrealistic assumption that a user would know or intuitively assume that the error formatting can be configured in order to get a more readable output to make matters worse often it might be really hard to reproduce an error due to the non idempotent nature of most applications so even if a user has figured out how to configure the error formatting correctly it might be a lot of work to reproduce the error solution i suggest to change the default error formatting from pretty to colorless in order to avoid the problem described above i m all in favor of trying to provide the best possible dx out of the box without further configuration but we need to do so in a safe way that avoids negative surprise once we ve changed the default error formatting i suggest to collaborate with framework authors e g redwood nexus so these frameworks enable pretty error formatting out of the box as they can guarantee a safe environment we should also investigate whether there are better heuristics to safely determine whether an environment support ascii colors
1
6,739
9,872,910,772
IssuesEvent
2019-06-22 09:21:46
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Both IDW interpolation and clip raster with mask indicate 100% complete, but hangs forever
Bug Feedback Processing
Author Name: **Tony Walters** (@tonywsea) Original Redmine Issue: [21660](https://issues.qgis.org/issues/21660) Affected QGIS version: 3.6.0 Redmine category:processing/qgis --- By clicking "cancel" many times eventually it will abort. Clean install of 3.6 on new Intel 8700 6 core system. Multi-core (6) rendering enabled. link to geopackage: https://we.tl/t-D6RQZztc9k
1.0
Both IDW interpolation and clip raster with mask indicate 100% complete, but hangs forever - Author Name: **Tony Walters** (@tonywsea) Original Redmine Issue: [21660](https://issues.qgis.org/issues/21660) Affected QGIS version: 3.6.0 Redmine category:processing/qgis --- By clicking "cancel" many times eventually it will abort. Clean install of 3.6 on new Intel 8700 6 core system. Multi-core (6) rendering enabled. link to geopackage: https://we.tl/t-D6RQZztc9k
process
both idw interpolation and clip raster with mask indicate complete but hangs forever author name tony walters tonywsea original redmine issue affected qgis version redmine category processing qgis by clicking cancel many times eventually it will abort clean install of on new intel core system multi core rendering enabled link to geopackage
1
15,411
19,599,455,233
IssuesEvent
2022-01-05 22:28:10
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Sign container images
enhancement process
### Problem Currently our container images are not signed, which puts us at risk for supply-chain attacks. There are [new tools](https://github.blog/2021-12-06-safeguard-container-signing-capability-actions/) that making container signing relatively trivial. ### Solution Generate a signing key via `cosign generate-key-pair` and store in a `COSIGN_KEY` GitHub repository secret. In `release` GitHub workflows: - Install cosign - Sign the container with annotations for repo and ref. ### Alternatives _No response_
1.0
Sign container images - ### Problem Currently our container images are not signed, which puts us at risk for supply-chain attacks. There are [new tools](https://github.blog/2021-12-06-safeguard-container-signing-capability-actions/) that making container signing relatively trivial. ### Solution Generate a signing key via `cosign generate-key-pair` and store in a `COSIGN_KEY` GitHub repository secret. In `release` GitHub workflows: - Install cosign - Sign the container with annotations for repo and ref. ### Alternatives _No response_
process
sign container images problem currently our container images are not signed which puts us at risk for supply chain attacks there are that making container signing relatively trivial solution generate a signing key via cosign generate key pair and store in a cosign key github repository secret in release github workflows install cosign sign the container with annotations for repo and ref alternatives no response
1
279,357
30,702,518,117
IssuesEvent
2023-07-27 01:37:05
nidhi7598/linux-3.0.35_CVE-2018-13405
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
closed
CVE-2013-4512 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed
Mend: dependency security vulnerability
## CVE-2013-4512 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Buffer overflow in the exitcode_proc_write function in arch/um/kernel/exitcode.c in the Linux kernel before 3.12 allows local users to cause a denial of service or possibly have unspecified other impact by leveraging root privileges for a write operation. <p>Publish Date: 2013-11-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-4512>CVE-2013-4512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-4512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-4512</a></p> <p>Release Date: 2013-11-12</p> <p>Fix Resolution: v3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-4512 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2013-4512 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/um/kernel/exitcode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Buffer overflow in the exitcode_proc_write function in arch/um/kernel/exitcode.c in the Linux kernel before 3.12 allows local users to cause a denial of service or possibly have unspecified other impact by leveraging root privileges for a write operation. <p>Publish Date: 2013-11-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-4512>CVE-2013-4512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-4512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-4512</a></p> <p>Release Date: 2013-11-12</p> <p>Fix Resolution: v3.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files arch um kernel exitcode c arch um kernel exitcode c arch um kernel exitcode c vulnerability details buffer overflow in the exitcode proc write function in arch um kernel exitcode c in the linux kernel before allows local users to cause a denial of service or possibly have unspecified other impact by leveraging root privileges for a write operation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
7,296
10,442,376,326
IssuesEvent
2019-09-18 12:58:05
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Log format error - XFF Field combining date
bug log-processing log/date/time format
I'm having a problem getting the log format to parse. I have an XFF field but Goaccess is ignoring the delimiter of the last IP in the XFF field and combining it with the next field which is the date field e.g: Token '127.0.0.1|18/Sep/2019' doesn't match specifier '%d' ...my log format is as follows: time-format %H:%M:%S date-format %d/%b/%Y log-format %s|%U|%^|~h{,"}|%d:%t %^|%v|%u|%R ...and here are some log entries: ``` 200|GET /wp-content/uploads/property_images/src_BESAAP5405_884_V0_48DC-1624x914.jpg HTTP/1.1|84.126.152.23|84.126.152.23, 84.126.152.23, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:05 +0200|www.example.com|Mozilla/5.0 (iPhone; CPU iPhone OS 11_4_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/11.0 Mobile/15E148 Safari/604.1|https://www.example.com/es/ventas-propiedad/villa-a-la-venta-en-el-toro-port-adriano-calvia-v-toro-436/ 200|GET /wp-content/uploads/property_images/src_BESAAP5405_892_V0_78F2-1624x914.jpg HTTP/1.1|84.126.152.23|84.126.152.23|18/Sep/2019:07:01:11 +0200|www.example.com|Mozilla/5.0 (iPhone; CPU iPhone OS 11_4_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/11.0 Mobile/15E148 Safari/604.1|https://www.example.com/es/ventas-propiedad/villa-a-la-venta-en-el-toro-port-adriano-calvia-v-toro-436/ 404|GET /test/ HTTP/1.1|212.92.121.17|212.92.121.17, 212.92.121.17, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:15 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /test/ HTTP/1.1|212.92.121.17|212.92.121.17|18/Sep/2019:07:01:15 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /main/ HTTP/1.1|141.101.76.184|212.92.121.17|18/Sep/2019:07:01:18 +0200|example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /main/ HTTP/1.1|212.92.121.17|212.92.121.17, 212.92.121.17, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:20 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /main/ HTTP/1.1|212.92.121.17|212.92.121.17|18/Sep/2019:07:01:20 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /site/ HTTP/1.1|141.101.76.238|212.92.121.17|18/Sep/2019:07:01:21 +0200|example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /de/immobilienverkaufe/villen-verkauf-mallorca/ HTTP/1.1|157.55.39.176|157.55.39.176|18/Sep/2019:07:01:23 +0200|www.example.com|Mozilla/5.0 (compatible; bingbot/2.0; +http://www.bing.com/bingbot.htm)|- ``` Could you advise where I'm going wring please? Many thanks.
1.0
Log format error - XFF Field combining date - I'm having a problem getting the log format to parse. I have an XFF field but Goaccess is ignoring the delimiter of the last IP in the XFF field and combining it with the next field which is the date field e.g: Token '127.0.0.1|18/Sep/2019' doesn't match specifier '%d' ...my log format is as follows: time-format %H:%M:%S date-format %d/%b/%Y log-format %s|%U|%^|~h{,"}|%d:%t %^|%v|%u|%R ...and here are some log entries: ``` 200|GET /wp-content/uploads/property_images/src_BESAAP5405_884_V0_48DC-1624x914.jpg HTTP/1.1|84.126.152.23|84.126.152.23, 84.126.152.23, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:05 +0200|www.example.com|Mozilla/5.0 (iPhone; CPU iPhone OS 11_4_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/11.0 Mobile/15E148 Safari/604.1|https://www.example.com/es/ventas-propiedad/villa-a-la-venta-en-el-toro-port-adriano-calvia-v-toro-436/ 200|GET /wp-content/uploads/property_images/src_BESAAP5405_892_V0_78F2-1624x914.jpg HTTP/1.1|84.126.152.23|84.126.152.23|18/Sep/2019:07:01:11 +0200|www.example.com|Mozilla/5.0 (iPhone; CPU iPhone OS 11_4_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/11.0 Mobile/15E148 Safari/604.1|https://www.example.com/es/ventas-propiedad/villa-a-la-venta-en-el-toro-port-adriano-calvia-v-toro-436/ 404|GET /test/ HTTP/1.1|212.92.121.17|212.92.121.17, 212.92.121.17, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:15 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /test/ HTTP/1.1|212.92.121.17|212.92.121.17|18/Sep/2019:07:01:15 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /main/ HTTP/1.1|141.101.76.184|212.92.121.17|18/Sep/2019:07:01:18 +0200|example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /main/ HTTP/1.1|212.92.121.17|212.92.121.17, 212.92.121.17, 127.0.0.1, 127.0.0.1|18/Sep/2019:07:01:20 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 404|GET /main/ HTTP/1.1|212.92.121.17|212.92.121.17|18/Sep/2019:07:01:20 +0200|www.example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /site/ HTTP/1.1|141.101.76.238|212.92.121.17|18/Sep/2019:07:01:21 +0200|example.com|Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0|- 301|GET /de/immobilienverkaufe/villen-verkauf-mallorca/ HTTP/1.1|157.55.39.176|157.55.39.176|18/Sep/2019:07:01:23 +0200|www.example.com|Mozilla/5.0 (compatible; bingbot/2.0; +http://www.bing.com/bingbot.htm)|- ``` Could you advise where I'm going wring please? Many thanks.
process
log format error xff field combining date i m having a problem getting the log format to parse i have an xff field but goaccess is ignoring the delimiter of the last ip in the xff field and combining it with the next field which is the date field e g token sep doesn t match specifier d my log format is as follows time format h m s date format d b y log format s u h d t v u r and here are some log entries get wp content uploads property images src jpg http sep iphone cpu iphone os like mac os x applewebkit khtml like gecko version mobile safari get wp content uploads property images src jpg http sep iphone cpu iphone os like mac os x applewebkit khtml like gecko version mobile safari get test http sep windows nt rv gecko firefox get test http sep windows nt rv gecko firefox get main http sep example com mozilla windows nt rv gecko firefox get main http sep windows nt rv gecko firefox get main http sep windows nt rv gecko firefox get site http sep example com mozilla windows nt rv gecko firefox get de immobilienverkaufe villen verkauf mallorca http sep compatible bingbot could you advise where i m going wring please many thanks
1
10,885
13,654,073,975
IssuesEvent
2020-09-27 15:42:35
amor71/LiuAlgoTrader
https://api.github.com/repos/amor71/LiuAlgoTrader
closed
TOML configuration file
in-process
add env variable to control tradeplan/minder configuration file location
1.0
TOML configuration file - add env variable to control tradeplan/minder configuration file location
process
toml configuration file add env variable to control tradeplan minder configuration file location
1
3,500
2,538,568,941
IssuesEvent
2015-01-27 08:19:55
newca12/gapt
https://api.github.com/repos/newca12/gapt
closed
Parse our XML format for first-order resolution proofs
2–5 stars Component-Parsing duplicate enhancement imported Priority-Medium
_From [fra...@gmail.com](https://code.google.com/u/108596877348066494139/) on February 01, 2011 11:01:51_ Our XML parsing code should be extended to be able to parse resolution proofs stored in our XML format (calculus "RES"), as for example in http://www.logic.at/ceres/examples/lattice/lattice-out.xml . _Original issue: http://code.google.com/p/gapt/issues/detail?id=105_
1.0
Parse our XML format for first-order resolution proofs - _From [fra...@gmail.com](https://code.google.com/u/108596877348066494139/) on February 01, 2011 11:01:51_ Our XML parsing code should be extended to be able to parse resolution proofs stored in our XML format (calculus "RES"), as for example in http://www.logic.at/ceres/examples/lattice/lattice-out.xml . _Original issue: http://code.google.com/p/gapt/issues/detail?id=105_
non_process
parse our xml format for first order resolution proofs from on february our xml parsing code should be extended to be able to parse resolution proofs stored in our xml format calculus res as for example in original issue
0
14,792
18,066,273,282
IssuesEvent
2021-09-20 19:34:19
googleapis/nodejs-contact-center-insights
https://api.github.com/repos/googleapis/nodejs-contact-center-insights
closed
Remove sample links from linkinator config
type: process samples api: contactcenterinsights
After all the code samples are submitted, we should remove their links from the ignore list of [linkinator.config.json](https://github.com/googleapis/nodejs-contact-center-insights/blob/main/linkinator.config.json). We previously had to add the code sample links here in order to pass the docs presubmit check (This check made sure that only links that exist are added to README files, but when we submitted the code samples, their corresponding GitHub links don't exist yet, so we had to do this workaround).
1.0
Remove sample links from linkinator config - After all the code samples are submitted, we should remove their links from the ignore list of [linkinator.config.json](https://github.com/googleapis/nodejs-contact-center-insights/blob/main/linkinator.config.json). We previously had to add the code sample links here in order to pass the docs presubmit check (This check made sure that only links that exist are added to README files, but when we submitted the code samples, their corresponding GitHub links don't exist yet, so we had to do this workaround).
process
remove sample links from linkinator config after all the code samples are submitted we should remove their links from the ignore list of we previously had to add the code sample links here in order to pass the docs presubmit check this check made sure that only links that exist are added to readme files but when we submitted the code samples their corresponding github links don t exist yet so we had to do this workaround
1
10,823
13,609,293,118
IssuesEvent
2020-09-23 04:51:01
googleapis/java-spanner-jdbc
https://api.github.com/repos/googleapis/java-spanner-jdbc
closed
Dependency Dashboard
api: spanner type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/major-mockito.version -->deps: update dependency org.mockito:mockito-core to v3 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/major-mockito.version -->deps: update dependency org.mockito:mockito-core to v3 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any deps update dependency org mockito mockito core to check this box to trigger a request for renovate to run again on this repository
1
8,094
11,271,683,052
IssuesEvent
2020-01-14 13:32:50
code4romania/expert-consultation-client
https://api.github.com/repos/code4romania/expert-consultation-client
closed
Upload new document to platform
angular document processing documents enhancement
As an admin user of the Legal Consultation platform I want to be able to import a new document into the platform. Importing the document will trigger the saving of a document metadata information, the trigger of the document breakdown logic. If the processing is succesful, the loading of the document should redirect to the view document breakdown page for the saved document. ![Incarca document](https://user-images.githubusercontent.com/15039873/58744981-1690d380-83ff-11e9-97e3-b32e88d50848.png)
1.0
Upload new document to platform - As an admin user of the Legal Consultation platform I want to be able to import a new document into the platform. Importing the document will trigger the saving of a document metadata information, the trigger of the document breakdown logic. If the processing is succesful, the loading of the document should redirect to the view document breakdown page for the saved document. ![Incarca document](https://user-images.githubusercontent.com/15039873/58744981-1690d380-83ff-11e9-97e3-b32e88d50848.png)
process
upload new document to platform as an admin user of the legal consultation platform i want to be able to import a new document into the platform importing the document will trigger the saving of a document metadata information the trigger of the document breakdown logic if the processing is succesful the loading of the document should redirect to the view document breakdown page for the saved document
1
731,930
25,237,625,517
IssuesEvent
2022-11-15 03:10:09
squaredproject/Entwined
https://api.github.com/repos/squaredproject/Entwined
closed
Shrub mapping problem
1 HighPriority - chromatik
We need a way algorithmically to pull out all of the pixels on an single level on a shrub, or all the pixels on a single branch (cluster). Given that all the shrubs are the same, and that the pixels are laid out by the shrub fixture creation tool, this should not be difficult to back figure. However, my current attempts have failed. Shrub test pattern should work.
1.0
Shrub mapping problem - We need a way algorithmically to pull out all of the pixels on an single level on a shrub, or all the pixels on a single branch (cluster). Given that all the shrubs are the same, and that the pixels are laid out by the shrub fixture creation tool, this should not be difficult to back figure. However, my current attempts have failed. Shrub test pattern should work.
non_process
shrub mapping problem we need a way algorithmically to pull out all of the pixels on an single level on a shrub or all the pixels on a single branch cluster given that all the shrubs are the same and that the pixels are laid out by the shrub fixture creation tool this should not be difficult to back figure however my current attempts have failed shrub test pattern should work
0
45,444
5,715,709,950
IssuesEvent
2017-04-19 13:43:09
AffiliateWP/AffiliateWP
https://api.github.com/repos/AffiliateWP/AffiliateWP
closed
Filtering by date range/affiliate in Referrals screen shows the "All" referral count
bug Has PR needs testing
Currently when filtering the Referrals screen by affiliate and by date range, the item count (number of referrals) does not update - instead it shows the total number as per the _all_ referrals count. Would it be possible to only show the number of referrals listed based on the selected search parameters?
1.0
Filtering by date range/affiliate in Referrals screen shows the "All" referral count - Currently when filtering the Referrals screen by affiliate and by date range, the item count (number of referrals) does not update - instead it shows the total number as per the _all_ referrals count. Would it be possible to only show the number of referrals listed based on the selected search parameters?
non_process
filtering by date range affiliate in referrals screen shows the all referral count currently when filtering the referrals screen by affiliate and by date range the item count number of referrals does not update instead it shows the total number as per the all referrals count would it be possible to only show the number of referrals listed based on the selected search parameters
0
2,216
5,059,449,011
IssuesEvent
2016-12-22 08:01:14
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
reopened
Backward incompatibility in edge protocol
process_wontfix type_bug
Observed while testing a dev build, very likely introduced with #203 (part of #146) . Either backward compatibility needs to be restored or the breakage needs to be documented / a new major version for the client lib introduced. @wimpers , @saelbrec : this needs to happen before releasing RC2.
1.0
Backward incompatibility in edge protocol - Observed while testing a dev build, very likely introduced with #203 (part of #146) . Either backward compatibility needs to be restored or the breakage needs to be documented / a new major version for the client lib introduced. @wimpers , @saelbrec : this needs to happen before releasing RC2.
process
backward incompatibility in edge protocol observed while testing a dev build very likely introduced with part of either backward compatibility needs to be restored or the breakage needs to be documented a new major version for the client lib introduced wimpers saelbrec this needs to happen before releasing
1
341
2,657,904,560
IssuesEvent
2015-03-18 12:37:08
OpenSRP/opensrp-server
https://api.github.com/repos/OpenSRP/opensrp-server
opened
Handling inputs other than form submissions e.g. mjivita submission, birth sms notifications (through rapidpro)
Functional Requirement High Priority
This is currently not supported therefore it needs to be considered while working on: Parent Issue #3
1.0
Handling inputs other than form submissions e.g. mjivita submission, birth sms notifications (through rapidpro) - This is currently not supported therefore it needs to be considered while working on: Parent Issue #3
non_process
handling inputs other than form submissions e g mjivita submission birth sms notifications through rapidpro this is currently not supported therefore it needs to be considered while working on parent issue
0
22,558
31,777,753,541
IssuesEvent
2023-09-12 15:21:32
IMAP-Science-Operations-Center/imap_processing
https://api.github.com/repos/IMAP-Science-Operations-Center/imap_processing
opened
[L5] Produce L1A CDF formatted data products for IMAP-Ultra
IMAP-Ultra L1 Requirement: Level 5 Untested Parent Req:SDC Execution of L1 Processing Software
### Summary of the L5 requirement The SDC needs to produce L1A CDF formatted data products for the IMAP-Lo instrument. The data products for L1A can be found in the IMAP-Lo algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks **Tasks** - [ ] Update dependency page with all input dependencies <link to page> - [ ] Create skeleton CDFs and send to SPDF - [ ] (Need to figure out how to break up issues. by data product?) - [ ] #94
1.0
[L5] Produce L1A CDF formatted data products for IMAP-Ultra - ### Summary of the L5 requirement The SDC needs to produce L1A CDF formatted data products for the IMAP-Lo instrument. The data products for L1A can be found in the IMAP-Lo algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks **Tasks** - [ ] Update dependency page with all input dependencies <link to page> - [ ] Create skeleton CDFs and send to SPDF - [ ] (Need to figure out how to break up issues. by data product?) - [ ] #94
process
produce cdf formatted data products for imap ultra summary of the requirement the sdc needs to produce cdf formatted data products for the imap lo instrument the data products for can be found in the imap lo algorithm document here parent requirement soc sdc sdc execution of processing software the sdc shall be capable of executing processing software to produce data products within seven days of receipt of all requisite data inputs requirement the sdc produces data products via instrument team supplied algorithms and sdc produced processing code soc test sit processing sit tasks tasks update dependency page with all input dependencies create skeleton cdfs and send to spdf need to figure out how to break up issues by data product
1
602,925
18,516,667,662
IssuesEvent
2021-10-20 10:53:22
GoogleCloudPlatform/nodejs-getting-started
https://api.github.com/repos/GoogleCloudPlatform/nodejs-getting-started
reopened
background: "after all" hook for "should get the correct response" failed
type: bug priority: p1 flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 555ab0f0dcd02f4e58955bcc6b6c8465f4838ef0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/55cb6df5-f315-48e4-b834-0f0d633a5d9c), [Sponge](http://sponge2/55cb6df5-f315-48e4-b834-0f0d633a5d9c) status: failed <details><summary>Test output</summary><br><pre>16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client.js:330:49) at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at process.nextTick (node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78) at process._tickCallback (internal/process/next_tick.js:61:11) Caused by: Error at CollectionReference._get (node_modules/@google-cloud/firestore/build/src/reference.js:1519:23) at CollectionReference.get (node_modules/@google-cloud/firestore/build/src/reference.js:1507:21) at Context.after (test/app.test.js:73:54)</pre></details>
1.0
background: "after all" hook for "should get the correct response" failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 555ab0f0dcd02f4e58955bcc6b6c8465f4838ef0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/55cb6df5-f315-48e4-b834-0f0d633a5d9c), [Sponge](http://sponge2/55cb6df5-f315-48e4-b834-0f0d633a5d9c) status: failed <details><summary>Test output</summary><br><pre>16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client.js:330:49) at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at process.nextTick (node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78) at process._tickCallback (internal/process/next_tick.js:61:11) Caused by: Error at CollectionReference._get (node_modules/@google-cloud/firestore/build/src/reference.js:1519:23) at CollectionReference.get (node_modules/@google-cloud/firestore/build/src/reference.js:1507:21) at Context.after (test/app.test.js:73:54)</pre></details>
non_process
background after all hook for should get the correct response failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus node modules grpc grpc js build src call js at object onreceivestatus node modules grpc grpc js build src client js at object onreceivestatus node modules grpc grpc js build src client interceptors js at process nexttick node modules grpc grpc js build src call stream js at process tickcallback internal process next tick js caused by error at collectionreference get node modules google cloud firestore build src reference js at collectionreference get node modules google cloud firestore build src reference js at context after test app test js
0
16,372
4,047,286,889
IssuesEvent
2016-05-23 04:00:39
alanplotko/Dash
https://api.github.com/repos/alanplotko/Dash
closed
Add guidelines and useful information for contributors
documentation
Add guidelines to help contributors understand issue labels and other information that they may find useful.
1.0
Add guidelines and useful information for contributors - Add guidelines to help contributors understand issue labels and other information that they may find useful.
non_process
add guidelines and useful information for contributors add guidelines to help contributors understand issue labels and other information that they may find useful
0
20,683
27,354,500,284
IssuesEvent
2023-02-27 11:55:12
camunda/issues
https://api.github.com/repos/camunda/issues
opened
BPMN Compensation Events + Activities
component:desktopModeler component:operate component:optimize component:tasklist component:webModeler component:zeebe-process-automation public feature-parity potential:8.3
> This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Support for BPMN Compensation Events + Activities ### User Problem We execute tasks in our processes that sometimes have to be canceled later under certain circumstances. Compensation Events+Activities is a very powerful construct for handling such situations. Without these BPMN symbols, users would have to add a lot of additional complexity to their models. Today these symbols are not supported by Camunda 8, whereas we support them in Camunda 7. ### User Stories We might want to consider multiple phases if this helps us shipping a first iteration faster: a) - As a Developer I can model and execute "Intermediate Throwing Compensation Event" - As a Developer I can trust that this will execute all defined compensation handlers according to BPMN specification (support for BPMN Compensation activity + Compensation Boundary Event) b) - As a Developer I can use Compensation Start Events c) - As a Developer I can use Compensation End Events ### Implementation Notes <!-- Notes to consider for implementation, for example: * In Cawemo we already have the capability to manage templates via the feature that we call “catalog” * What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown * We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands. * See these high level wireframes to capture the idea --> ### Breakdown > This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} #### Implement phase ## <!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. --> #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral - See below comment - [Bernds Book](https://jira.camunda.com/browse/SUPPORT-9757) - [Zeebe](https://github.com/camunda/zeebe/issues/5134)
1.0
BPMN Compensation Events + Activities - > This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Support for BPMN Compensation Events + Activities ### User Problem We execute tasks in our processes that sometimes have to be canceled later under certain circumstances. Compensation Events+Activities is a very powerful construct for handling such situations. Without these BPMN symbols, users would have to add a lot of additional complexity to their models. Today these symbols are not supported by Camunda 8, whereas we support them in Camunda 7. ### User Stories We might want to consider multiple phases if this helps us shipping a first iteration faster: a) - As a Developer I can model and execute "Intermediate Throwing Compensation Event" - As a Developer I can trust that this will execute all defined compensation handlers according to BPMN specification (support for BPMN Compensation activity + Compensation Boundary Event) b) - As a Developer I can use Compensation Start Events c) - As a Developer I can use Compensation End Events ### Implementation Notes <!-- Notes to consider for implementation, for example: * In Cawemo we already have the capability to manage templates via the feature that we call “catalog” * What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown * We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands. * See these high level wireframes to capture the idea --> ### Breakdown > This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} #### Implement phase ## <!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. --> #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral - See below comment - [Bernds Book](https://jira.camunda.com/browse/SUPPORT-9757) - [Zeebe](https://github.com/camunda/zeebe/issues/5134)
process
bpmn compensation events activities this is an epic internal docs issue it bundles all activities we conduct around a certain initiative it will typically links to various child issues from various repositories and can be spread across multiple teams value proposition statement support for bpmn compensation events activities user problem we execute tasks in our processes that sometimes have to be canceled later under certain circumstances compensation events activities is a very powerful construct for handling such situations without these bpmn symbols users would have to add a lot of additional complexity to their models today these symbols are not supported by camunda whereas we support them in camunda user stories we might want to consider multiple phases if this helps us shipping a first iteration faster a as a developer i can model and execute intermediate throwing compensation event as a developer i can trust that this will execute all defined compensation handlers according to bpmn specification support for bpmn compensation activity compensation boundary event b as a developer i can use compensation start events c as a developer i can use compensation end events implementation notes notes to consider for implementation for example in cawemo we already have the capability to manage templates via the feature that we call “catalog” what we would build now is the ability to a use this feature in the web modeler to create templates and b when the context pad opens for defining the type of a task the templates that decorate service tasks are shown we should clarify terminology integrations vs connectors vs job workers vs element templates particularly “element templates” might not be a term that a user intuitively understands see these high level wireframes to capture the idea breakdown this section links to various sub issues tasks contributing to respective epic phase or phase results where appropriate discovery phase define phase design planning reviewed by design date designer assigned yes no design necessary or no designer available assignee design brief link to design brief research brief link to research brief design deliverables deliverable name link to gh issue implement phase validate phase links to additional collateral see below comment
1
1,462
4,044,092,529
IssuesEvent
2016-05-21 03:50:28
nodejs/node
https://api.github.com/repos/nodejs/node
opened
doc: formatting issues for child_process docs
child_process doc good first contribution
* **Version**: v5.11.0+ * **Platform**: n/a * **Subsystem**: doc, child_process There are some formatting issues for `exec()` and `execFile()` in the `child_process` docs. Specifically: * f85412d49b0ac1b3d2fa1e9c0dfe2491ea3aa9be erroneously added a backslash in front of an asterisk that was already contained within backticks. This causes the backslash to show up when rendered. The backslash can be safely removed. * The `maxBuffer` link is not styled like the other properties listed and looks out of place. My suggestion would be to have `<code>`-wrapped content within links to continue to have the same gray background color normally, but switch to the green background color on hover like normal links. There may be other instances of this in the docs, I have not checked.
1.0
doc: formatting issues for child_process docs - * **Version**: v5.11.0+ * **Platform**: n/a * **Subsystem**: doc, child_process There are some formatting issues for `exec()` and `execFile()` in the `child_process` docs. Specifically: * f85412d49b0ac1b3d2fa1e9c0dfe2491ea3aa9be erroneously added a backslash in front of an asterisk that was already contained within backticks. This causes the backslash to show up when rendered. The backslash can be safely removed. * The `maxBuffer` link is not styled like the other properties listed and looks out of place. My suggestion would be to have `<code>`-wrapped content within links to continue to have the same gray background color normally, but switch to the green background color on hover like normal links. There may be other instances of this in the docs, I have not checked.
process
doc formatting issues for child process docs version platform n a subsystem doc child process there are some formatting issues for exec and execfile in the child process docs specifically erroneously added a backslash in front of an asterisk that was already contained within backticks this causes the backslash to show up when rendered the backslash can be safely removed the maxbuffer link is not styled like the other properties listed and looks out of place my suggestion would be to have wrapped content within links to continue to have the same gray background color normally but switch to the green background color on hover like normal links there may be other instances of this in the docs i have not checked
1
17,150
22,700,803,923
IssuesEvent
2022-07-05 10:27:38
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Output mapping doesn't create an incident if a variable is missing
kind/bug scope/broker severity/mid area/reliability team/process-automation
**Describe the bug** I deployed a BPMN process with a task. The task has an output mapping that adds (or updates) a property of a variable `a.b`. ![image](https://user-images.githubusercontent.com/4305769/173822959-aaef47ec-7259-473e-ae9c-5a0302942b3a.png) Now, the job of the task is completed (accidentally) without the variable `x` that is referenced as the variable assignment value. If the process instance doesn't have the variable `a` then an incident is created, as expected. The incident has the following error message: ``` failed to evaluate expression '{a:if (a = null) then {b: x} else put all(a,{b: x})}': no variable found for name 'x' ``` But if the process instance has the variable `a` then no incident is created. Instead, the variable `a` is set to `null` and the task is completed. ![image](https://user-images.githubusercontent.com/4305769/173825201-68e6edcf-63d6-4d56-9d28-9aad9d624990.png) **To Reproduce** 1. Deploy the BPMN process: [output_mapping.bpmn.txt](https://github.com/camunda/zeebe/files/8909195/output_mapping.bpmn.txt) (without the `.txt` ending) 2. Create a new process instance with the variables `{ "a": { "c": 1 } }` 3. Complete the task with job type `task-a` without variables 4. Verify that no incident is created and the task is completed **Expected behavior** An incident is created if a job is completed without a variable that is referenced as the variable assignment value of an output mapping. The behavior should be consistent. For the case that the variable is present and the case that the variable is not present. **Log/Stacktrace** None. **Environment:** - OS: - Zeebe Version: `8.0.2` - Configuration:
1.0
Output mapping doesn't create an incident if a variable is missing - **Describe the bug** I deployed a BPMN process with a task. The task has an output mapping that adds (or updates) a property of a variable `a.b`. ![image](https://user-images.githubusercontent.com/4305769/173822959-aaef47ec-7259-473e-ae9c-5a0302942b3a.png) Now, the job of the task is completed (accidentally) without the variable `x` that is referenced as the variable assignment value. If the process instance doesn't have the variable `a` then an incident is created, as expected. The incident has the following error message: ``` failed to evaluate expression '{a:if (a = null) then {b: x} else put all(a,{b: x})}': no variable found for name 'x' ``` But if the process instance has the variable `a` then no incident is created. Instead, the variable `a` is set to `null` and the task is completed. ![image](https://user-images.githubusercontent.com/4305769/173825201-68e6edcf-63d6-4d56-9d28-9aad9d624990.png) **To Reproduce** 1. Deploy the BPMN process: [output_mapping.bpmn.txt](https://github.com/camunda/zeebe/files/8909195/output_mapping.bpmn.txt) (without the `.txt` ending) 2. Create a new process instance with the variables `{ "a": { "c": 1 } }` 3. Complete the task with job type `task-a` without variables 4. Verify that no incident is created and the task is completed **Expected behavior** An incident is created if a job is completed without a variable that is referenced as the variable assignment value of an output mapping. The behavior should be consistent. For the case that the variable is present and the case that the variable is not present. **Log/Stacktrace** None. **Environment:** - OS: - Zeebe Version: `8.0.2` - Configuration:
process
output mapping doesn t create an incident if a variable is missing describe the bug i deployed a bpmn process with a task the task has an output mapping that adds or updates a property of a variable a b now the job of the task is completed accidentally without the variable x that is referenced as the variable assignment value if the process instance doesn t have the variable a then an incident is created as expected the incident has the following error message failed to evaluate expression a if a null then b x else put all a b x no variable found for name x but if the process instance has the variable a then no incident is created instead the variable a is set to null and the task is completed to reproduce deploy the bpmn process without the txt ending create a new process instance with the variables a c complete the task with job type task a without variables verify that no incident is created and the task is completed expected behavior an incident is created if a job is completed without a variable that is referenced as the variable assignment value of an output mapping the behavior should be consistent for the case that the variable is present and the case that the variable is not present log stacktrace none environment os zeebe version configuration
1
87,468
10,546,555,739
IssuesEvent
2019-10-02 21:48:08
vtex-apps/io-documentation
https://api.github.com/repos/vtex-apps/io-documentation
closed
vtex-apps/ideacrm has no documentation yet
no-documentation
[vtex-apps/ideacrm](https://github.com/vtex-apps/ideacrm) hasn't created any README file yet or is not using Docs Builder
1.0
vtex-apps/ideacrm has no documentation yet - [vtex-apps/ideacrm](https://github.com/vtex-apps/ideacrm) hasn't created any README file yet or is not using Docs Builder
non_process
vtex apps ideacrm has no documentation yet hasn t created any readme file yet or is not using docs builder
0
36,191
2,796,912,873
IssuesEvent
2015-05-12 10:34:55
bioinformatics-ua/catalogue
https://api.github.com/repos/bioinformatics-ua/catalogue
closed
Footer: copyright of 404 and 500 page
enhancement normal priority
They have wrong year of copyright. More information should appear. This could be implemented against master branch.
1.0
Footer: copyright of 404 and 500 page - They have wrong year of copyright. More information should appear. This could be implemented against master branch.
non_process
footer copyright of and page they have wrong year of copyright more information should appear this could be implemented against master branch
0
18,388
24,518,182,742
IssuesEvent
2022-10-11 07:26:53
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[BUG] 我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务假死。等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况
compute/data-processing type/bug
### 问题简述: 我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务部分checkpoint失败,等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况。 另外,StartCursor除了指定latest可以正常消费外。使用fromMessageTime、earliest,都无法正常消费,但是可以在pulsar中看到,subscription是被正确创建出来了的。 ### 集群信息: pulsar版本:2.8.2 flink版本:1.14.4 connector版本: source:flink官方1.14.4 sink:sn-1.14.3.4 订阅模式:Key-shard ### 操作步骤 我对pulsar集群进行了broker扩容,扩容后进行unload操作,但因为新上节点未配置域名解析,导致flink集群无法正确解析pulsar新broker节点的域名,报错unknown host。 我于是配置了域名解析,但是flink的pulsar消费程序并没有死掉,只是卡住了。 我第二天发现了这个问题,于是kill了这个flink程序并进行了重启操作。 重启后,flink程序仍然无法正确消费这个topic的数据。 随后我又进行了一系列操作,比如不从checkpoint恢复,重置了cursor点位到某个ts时间点进行消费,发现仍然无法消费,但是通过pulsar监控发现,subscription的消费点位重置是成功的。 另外我还尝试了在上述操作的基础上修改订阅类型为Exclusive,仍然无法消费。 随后我更换了subscriptionId,尝试重置cursor点位到某个ts时间点进行消费,发现新的subscriptionId在pulsar创建成功,cursor位置正常,但仍无法消费。 最终我再次更换了subscriptionId,尝试使用latest进行消费,消费恢复正常。 ### 部分代码 [ProcessX01SignalDataTask.java.zip](https://github.com/streamnative/flink/files/9502458/ProcessX01SignalDataTask.java.zip) ### 部分日志 扩容broker未配置域名解析导致的报错假死日志: [报错假死.txt](https://github.com/streamnative/flink/files/9502468/default.txt) 重启后无法正常消费的日志: [重启无法正常消费.txt](https://github.com/streamnative/flink/files/9502602/default.txt) 一些其他尝试中的某次日志(具体不记得了): [一些其他尝试.txt](https://github.com/streamnative/flink/files/9502611/default.txt)
1.0
[BUG] 我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务假死。等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况 - ### 问题简述: 我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务部分checkpoint失败,等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况。 另外,StartCursor除了指定latest可以正常消费外。使用fromMessageTime、earliest,都无法正常消费,但是可以在pulsar中看到,subscription是被正确创建出来了的。 ### 集群信息: pulsar版本:2.8.2 flink版本:1.14.4 connector版本: source:flink官方1.14.4 sink:sn-1.14.3.4 订阅模式:Key-shard ### 操作步骤 我对pulsar集群进行了broker扩容,扩容后进行unload操作,但因为新上节点未配置域名解析,导致flink集群无法正确解析pulsar新broker节点的域名,报错unknown host。 我于是配置了域名解析,但是flink的pulsar消费程序并没有死掉,只是卡住了。 我第二天发现了这个问题,于是kill了这个flink程序并进行了重启操作。 重启后,flink程序仍然无法正确消费这个topic的数据。 随后我又进行了一系列操作,比如不从checkpoint恢复,重置了cursor点位到某个ts时间点进行消费,发现仍然无法消费,但是通过pulsar监控发现,subscription的消费点位重置是成功的。 另外我还尝试了在上述操作的基础上修改订阅类型为Exclusive,仍然无法消费。 随后我更换了subscriptionId,尝试重置cursor点位到某个ts时间点进行消费,发现新的subscriptionId在pulsar创建成功,cursor位置正常,但仍无法消费。 最终我再次更换了subscriptionId,尝试使用latest进行消费,消费恢复正常。 ### 部分代码 [ProcessX01SignalDataTask.java.zip](https://github.com/streamnative/flink/files/9502458/ProcessX01SignalDataTask.java.zip) ### 部分日志 扩容broker未配置域名解析导致的报错假死日志: [报错假死.txt](https://github.com/streamnative/flink/files/9502468/default.txt) 重启后无法正常消费的日志: [重启无法正常消费.txt](https://github.com/streamnative/flink/files/9502602/default.txt) 一些其他尝试中的某次日志(具体不记得了): [一些其他尝试.txt](https://github.com/streamnative/flink/files/9502611/default.txt)
process
我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务假死。等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况 问题简述: 我用flink消费pulsar数据,当pulsar集群不稳定出现异常之后,会导致flink任务部分checkpoint失败,等pulsar集群稳定后,重启flink消费程序,会出现无法消费的情况。 另外,startcursor除了指定latest可以正常消费外。使用frommessagetime、earliest,都无法正常消费,但是可以在pulsar中看到,subscription是被正确创建出来了的。 集群信息: pulsar版本: flink版本: connector版本: source: sink:sn 订阅模式:key shard 操作步骤 我对pulsar集群进行了broker扩容,扩容后进行unload操作,但因为新上节点未配置域名解析,导致flink集群无法正确解析pulsar新broker节点的域名,报错unknown host。 我于是配置了域名解析,但是flink的pulsar消费程序并没有死掉,只是卡住了。 我第二天发现了这个问题,于是kill了这个flink程序并进行了重启操作。 重启后,flink程序仍然无法正确消费这个topic的数据。 随后我又进行了一系列操作,比如不从checkpoint恢复,重置了cursor点位到某个ts时间点进行消费,发现仍然无法消费,但是通过pulsar监控发现,subscription的消费点位重置是成功的。 另外我还尝试了在上述操作的基础上修改订阅类型为exclusive,仍然无法消费。 随后我更换了subscriptionid,尝试重置cursor点位到某个ts时间点进行消费,发现新的subscriptionid在pulsar创建成功,cursor位置正常,但仍无法消费。 最终我再次更换了subscriptionid,尝试使用latest进行消费,消费恢复正常。 部分代码 部分日志 扩容broker未配置域名解析导致的报错假死日志: 重启后无法正常消费的日志: 一些其他尝试中的某次日志(具体不记得了):
1
110,570
11,705,704,083
IssuesEvent
2020-03-07 17:29:46
ExecutableBookProject/MyST-Parser
https://api.github.com/repos/ExecutableBookProject/MyST-Parser
closed
A way to programmatically generate docs for syntax additions?
documentation
I'm loving all of the documentation improvements that we've made on this. It helps me to wrap my head around the tool and think about how it extends commonmark markdown. One thing I'm concerned with is that over time we forget to clearly document the departures that we've made from CommonMark. Perhaps there is a way that we can programmatically generate a "reference spec" that can be inserted into the docs, so we know there is always a canonical source of truth for what *should* be in the documentation?
1.0
A way to programmatically generate docs for syntax additions? - I'm loving all of the documentation improvements that we've made on this. It helps me to wrap my head around the tool and think about how it extends commonmark markdown. One thing I'm concerned with is that over time we forget to clearly document the departures that we've made from CommonMark. Perhaps there is a way that we can programmatically generate a "reference spec" that can be inserted into the docs, so we know there is always a canonical source of truth for what *should* be in the documentation?
non_process
a way to programmatically generate docs for syntax additions i m loving all of the documentation improvements that we ve made on this it helps me to wrap my head around the tool and think about how it extends commonmark markdown one thing i m concerned with is that over time we forget to clearly document the departures that we ve made from commonmark perhaps there is a way that we can programmatically generate a reference spec that can be inserted into the docs so we know there is always a canonical source of truth for what should be in the documentation
0
820,101
30,759,491,396
IssuesEvent
2023-07-29 14:00:22
Gmu-Wiki/Gmu-Wiki-Back
https://api.github.com/repos/Gmu-Wiki/Gmu-Wiki-Back
closed
문의글 승인 api 수정 및 리팩터링
:zero: Priority: Critical :recycle: Refactor :ambulance: Hotfix :zap: Simple :question: Invalid
- 문의글 승인 부분에서 글의 inquiryName 을 가져와 반환하지 않고 "문의글 제목" 이라는 글을 담아 메일을 발송하고 있었음 - 더 이상 사용되지 않는 코드 삭제
1.0
문의글 승인 api 수정 및 리팩터링 - - 문의글 승인 부분에서 글의 inquiryName 을 가져와 반환하지 않고 "문의글 제목" 이라는 글을 담아 메일을 발송하고 있었음 - 더 이상 사용되지 않는 코드 삭제
non_process
문의글 승인 api 수정 및 리팩터링 문의글 승인 부분에서 글의 inquiryname 을 가져와 반환하지 않고 문의글 제목 이라는 글을 담아 메일을 발송하고 있었음 더 이상 사용되지 않는 코드 삭제
0
15,769
19,914,611,738
IssuesEvent
2022-01-25 21:00:54
plazi/community
https://api.github.com/repos/plazi/community
opened
to be processed 10.1093/isd/ixab026
process request
to be processed the file is too big to upload, here is the link https://academic.oup.com/isd/article-pdf/6/1/5/42268871/ixab026.pdf or https://doi.org/10.1093/isd/ixab026 genus level, so no material citation. Should show up in GBIF
1.0
to be processed 10.1093/isd/ixab026 - to be processed the file is too big to upload, here is the link https://academic.oup.com/isd/article-pdf/6/1/5/42268871/ixab026.pdf or https://doi.org/10.1093/isd/ixab026 genus level, so no material citation. Should show up in GBIF
process
to be processed isd to be processed the file is too big to upload here is the link or genus level so no material citation should show up in gbif
1
15,478
19,687,452,583
IssuesEvent
2022-01-12 00:33:07
crim-ca/weaver
https://api.github.com/repos/crim-ca/weaver
closed
CWL workflow outputs within nested binding not handled
triage/bug process/workflow
As presented in: https://github.com/crim-ca/weaver/issues/358#issuecomment-977194614 Workflows that define nested directories such as `outputBindings: {glob: "somedir/*.patterm"}` do not find outputs for mapping from `step(i) out ->step(i+1) in` because the `somedir` hierarchy is not preserved since results are staged-in for remote directly under job UUID. Sample definitions to test: 1. There is an application that generates the images ```json { "processDescription": { "process": { "visibility": "public", "id": "test_generation", "title": "Test image generation", "abstract": "Generates some test GeoTIFF", "version": "0.0.1", "outputs": [ { "id": "output_tifs", "title": "GeoTIFF Images", "formats": [ { "default": true, "mimeType": "image/tiff" } ] } ] }, "processVersion": "2.0" }, "immediateDeployment": true, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "CommandLineTool", "baseCommand": [ "generate", "--output", "output/" ], "requirements": { "DockerRequirement": { "dockerPull": "image_utils:latest" } }, "inputs": { "base_name": { "type": "string", "inputBinding": { "position": 1, "prefix": "--base-name" } }, "count": { "type": "int", "inputBinding": { "position": 2, "prefix": "--count" } } }, "outputs": { "output_tifs": { "type": { "type": "array", "items": "File" }, "outputBinding": { "glob": "output/*.tif" } } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/dockerizedApplication" } ``` 2. There is an application that blurs the images. ```json { "processDescription": { "process": { "visibility": "public", "id": "test_blurring", "title": "Test image blurring", "abstract": "Test image blurring", "version": "0.0.1", "inputs": [ { "id": "input_files", "title": "Input images", "formats": [ { "mimeType": "image/tiff", "default": true } ], "minOccurs": "1", "maxOccurs": "unbounded" } ], "outputs": [ { "id": "output_tifs", "title": "Blurred images", "formats": [ { "mimeType": "image/tiff", "default": true } ] } ] }, "processVersion": "2.0" }, "immediateDeployment": true, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "CommandLineTool", "baseCommand": [ "blur", "-o", "output/" ], "requirements": { "DockerRequirement": { "dockerPull": "image_utils:latest" } }, "inputs": { "input_files": { "type": { "type": "array", "items": "File" }, "inputBinding": { "position": 1, "prefix": "--image" } } }, "outputs": { "output_tifs": { "type": { "type": "array", "items": "File" }, "outputBinding": { "glob": "output/*.tif" } } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/dockerizedApplication" } ``` 3. Register the workflow ```json { "processDescription": { "process": { "visibility": "public", "id": "test_workflow", "title": "Workflow to test Weaver", "abstract": "Workflow to test Weaver", "version": "0.0.1" } }, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "Workflow", "inputs": { "base_name": "string", "count": "int" }, "outputs": { "output": { "type": { "type": "array", "items": "File" }, "outputSource": "blurring/output_tifs" } }, "steps": { "generation": { "run": "test_generation", "in": { "base_name": "base_name", "count": "count" }, "out": [ "output_tifs" ] }, "blurring": { "run": "test_blurring", "in": { "input_files": "generation/output_tifs" }, "out": [ "output_tifs" ] } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/workflow" } ``` 4. Execute the workflow ```json { "mode": "async", "response": "document", "inputs": [ { "id": "base_name", "data": "castellon" }, { "id": "count", "data": 4 } ], "outputs": [ { "id": "output", "transmissionMode": "reference" } ] } ``` Again, the workflow is failing with the following error: ``` gip-ems-worker | [2021-11-23 21:38:07,374] INFO [MainThread][weaver.processes.execution] 00:00:00 5% accepted Following updates could take a while until the Application Package answers... gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][PYWPS] Missing parameter value: input_files gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][PYWPS] Exception: code: 400, description: input_files, locator: input_files gip-ems-worker | NoneType: None gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][weaver.processes.execution] Failed running [Job <0f8d2d9f-2826-4bd3-8d10-c643bcda91d2>] gip-ems-worker | Traceback (most recent call last): gip-ems-worker | File "/opt/local/src/weaver/weaver/processes/execution.py", line 165, in execute_process gip-ems-worker | mode=mode, job_uuid=job.id, remote_process=process) gip-ems-worker | File "/opt/local/src/weaver/weaver/wps/service.py", line 271, in execute_job gip-ems-worker | wps_response = super(WorkerService, self).execute(worker_process_id, wps_request, job_uuid) gip-ems-worker | File "/usr/local/lib/python3.7/site-packages/pywps/app/Service.py", line 79, in execute gip-ems-worker | return self._parse_and_execute(process, wps_request, uuid) gip-ems-worker | File "/usr/local/lib/python3.7/site-packages/pywps/app/Service.py", line 136, in _parse_and_execute gip-ems-worker | inpt.identifier, inpt.identifier) gip-ems-worker | pywps.exceptions.MissingParameterValue: 400 MissingParameterValue: input_files ``` _Originally posted by @lhcorralo in https://github.com/crim-ca/weaver/issues/358#issuecomment-977194614_
1.0
CWL workflow outputs within nested binding not handled - As presented in: https://github.com/crim-ca/weaver/issues/358#issuecomment-977194614 Workflows that define nested directories such as `outputBindings: {glob: "somedir/*.patterm"}` do not find outputs for mapping from `step(i) out ->step(i+1) in` because the `somedir` hierarchy is not preserved since results are staged-in for remote directly under job UUID. Sample definitions to test: 1. There is an application that generates the images ```json { "processDescription": { "process": { "visibility": "public", "id": "test_generation", "title": "Test image generation", "abstract": "Generates some test GeoTIFF", "version": "0.0.1", "outputs": [ { "id": "output_tifs", "title": "GeoTIFF Images", "formats": [ { "default": true, "mimeType": "image/tiff" } ] } ] }, "processVersion": "2.0" }, "immediateDeployment": true, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "CommandLineTool", "baseCommand": [ "generate", "--output", "output/" ], "requirements": { "DockerRequirement": { "dockerPull": "image_utils:latest" } }, "inputs": { "base_name": { "type": "string", "inputBinding": { "position": 1, "prefix": "--base-name" } }, "count": { "type": "int", "inputBinding": { "position": 2, "prefix": "--count" } } }, "outputs": { "output_tifs": { "type": { "type": "array", "items": "File" }, "outputBinding": { "glob": "output/*.tif" } } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/dockerizedApplication" } ``` 2. There is an application that blurs the images. ```json { "processDescription": { "process": { "visibility": "public", "id": "test_blurring", "title": "Test image blurring", "abstract": "Test image blurring", "version": "0.0.1", "inputs": [ { "id": "input_files", "title": "Input images", "formats": [ { "mimeType": "image/tiff", "default": true } ], "minOccurs": "1", "maxOccurs": "unbounded" } ], "outputs": [ { "id": "output_tifs", "title": "Blurred images", "formats": [ { "mimeType": "image/tiff", "default": true } ] } ] }, "processVersion": "2.0" }, "immediateDeployment": true, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "CommandLineTool", "baseCommand": [ "blur", "-o", "output/" ], "requirements": { "DockerRequirement": { "dockerPull": "image_utils:latest" } }, "inputs": { "input_files": { "type": { "type": "array", "items": "File" }, "inputBinding": { "position": 1, "prefix": "--image" } } }, "outputs": { "output_tifs": { "type": { "type": "array", "items": "File" }, "outputBinding": { "glob": "output/*.tif" } } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/dockerizedApplication" } ``` 3. Register the workflow ```json { "processDescription": { "process": { "visibility": "public", "id": "test_workflow", "title": "Workflow to test Weaver", "abstract": "Workflow to test Weaver", "version": "0.0.1" } }, "executionUnit": [ { "unit": { "cwlVersion": "v1.0", "class": "Workflow", "inputs": { "base_name": "string", "count": "int" }, "outputs": { "output": { "type": { "type": "array", "items": "File" }, "outputSource": "blurring/output_tifs" } }, "steps": { "generation": { "run": "test_generation", "in": { "base_name": "base_name", "count": "count" }, "out": [ "output_tifs" ] }, "blurring": { "run": "test_blurring", "in": { "input_files": "generation/output_tifs" }, "out": [ "output_tifs" ] } } } } ], "deploymentProfileName": "http://www.opengis.net/profiles/eoc/workflow" } ``` 4. Execute the workflow ```json { "mode": "async", "response": "document", "inputs": [ { "id": "base_name", "data": "castellon" }, { "id": "count", "data": 4 } ], "outputs": [ { "id": "output", "transmissionMode": "reference" } ] } ``` Again, the workflow is failing with the following error: ``` gip-ems-worker | [2021-11-23 21:38:07,374] INFO [MainThread][weaver.processes.execution] 00:00:00 5% accepted Following updates could take a while until the Application Package answers... gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][PYWPS] Missing parameter value: input_files gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][PYWPS] Exception: code: 400, description: input_files, locator: input_files gip-ems-worker | NoneType: None gip-ems-worker | [2021-11-23 21:38:07,383] ERROR [MainThread][weaver.processes.execution] Failed running [Job <0f8d2d9f-2826-4bd3-8d10-c643bcda91d2>] gip-ems-worker | Traceback (most recent call last): gip-ems-worker | File "/opt/local/src/weaver/weaver/processes/execution.py", line 165, in execute_process gip-ems-worker | mode=mode, job_uuid=job.id, remote_process=process) gip-ems-worker | File "/opt/local/src/weaver/weaver/wps/service.py", line 271, in execute_job gip-ems-worker | wps_response = super(WorkerService, self).execute(worker_process_id, wps_request, job_uuid) gip-ems-worker | File "/usr/local/lib/python3.7/site-packages/pywps/app/Service.py", line 79, in execute gip-ems-worker | return self._parse_and_execute(process, wps_request, uuid) gip-ems-worker | File "/usr/local/lib/python3.7/site-packages/pywps/app/Service.py", line 136, in _parse_and_execute gip-ems-worker | inpt.identifier, inpt.identifier) gip-ems-worker | pywps.exceptions.MissingParameterValue: 400 MissingParameterValue: input_files ``` _Originally posted by @lhcorralo in https://github.com/crim-ca/weaver/issues/358#issuecomment-977194614_
process
cwl workflow outputs within nested binding not handled as presented in workflows that define nested directories such as outputbindings glob somedir patterm do not find outputs for mapping from step i out step i in because the somedir hierarchy is not preserved since results are staged in for remote directly under job uuid sample definitions to test there is an application that generates the images json processdescription process visibility public id test generation title test image generation abstract generates some test geotiff version outputs id output tifs title geotiff images formats default true mimetype image tiff processversion immediatedeployment true executionunit unit cwlversion class commandlinetool basecommand generate output output requirements dockerrequirement dockerpull image utils latest inputs base name type string inputbinding position prefix base name count type int inputbinding position prefix count outputs output tifs type type array items file outputbinding glob output tif deploymentprofilename there is an application that blurs the images json processdescription process visibility public id test blurring title test image blurring abstract test image blurring version inputs id input files title input images formats mimetype image tiff default true minoccurs maxoccurs unbounded outputs id output tifs title blurred images formats mimetype image tiff default true processversion immediatedeployment true executionunit unit cwlversion class commandlinetool basecommand blur o output requirements dockerrequirement dockerpull image utils latest inputs input files type type array items file inputbinding position prefix image outputs output tifs type type array items file outputbinding glob output tif deploymentprofilename register the workflow json processdescription process visibility public id test workflow title workflow to test weaver abstract workflow to test weaver version executionunit unit cwlversion class workflow inputs base name string count int outputs output type type array items file outputsource blurring output tifs steps generation run test generation in base name base name count count out output tifs blurring run test blurring in input files generation output tifs out output tifs deploymentprofilename execute the workflow json mode async response document inputs id base name data castellon id count data outputs id output transmissionmode reference again the workflow is failing with the following error gip ems worker info accepted following updates could take a while until the application package answers gip ems worker error missing parameter value input files gip ems worker error exception code description input files locator input files gip ems worker nonetype none gip ems worker error failed running gip ems worker traceback most recent call last gip ems worker file opt local src weaver weaver processes execution py line in execute process gip ems worker mode mode job uuid job id remote process process gip ems worker file opt local src weaver weaver wps service py line in execute job gip ems worker wps response super workerservice self execute worker process id wps request job uuid gip ems worker file usr local lib site packages pywps app service py line in execute gip ems worker return self parse and execute process wps request uuid gip ems worker file usr local lib site packages pywps app service py line in parse and execute gip ems worker inpt identifier inpt identifier gip ems worker pywps exceptions missingparametervalue missingparametervalue input files originally posted by lhcorralo in
1
193,675
22,216,261,586
IssuesEvent
2022-06-08 02:12:12
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2018-19985 (Medium) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2018-19985 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/hso.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/hso.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function hso_get_config_data in drivers/net/usb/hso.c in the Linux kernel through 4.19.8 reads if_num from the USB device (as a u8) and uses it to index a small array, resulting in an object out-of-bounds (OOB) read that potentially allows arbitrary read in the kernel address space. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19985>CVE-2018-19985</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-19985">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-19985</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 4.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19985 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2018-19985 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/hso.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/hso.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function hso_get_config_data in drivers/net/usb/hso.c in the Linux kernel through 4.19.8 reads if_num from the USB device (as a u8) and uses it to index a small array, resulting in an object out-of-bounds (OOB) read that potentially allows arbitrary read in the kernel address space. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19985>CVE-2018-19985</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-19985">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-19985</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 4.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers net usb hso c drivers net usb hso c vulnerability details the function hso get config data in drivers net usb hso c in the linux kernel through reads if num from the usb device as a and uses it to index a small array resulting in an object out of bounds oob read that potentially allows arbitrary read in the kernel address space publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
836
3,298,778,833
IssuesEvent
2015-11-02 15:56:43
nehbit/aether-public
https://api.github.com/repos/nehbit/aether-public
closed
Release qt5reactor separately
discussion process
Would be great if the `qt5reactor.py` you folks have written could be broken out and put in a separate repo, packaged using setuptools and put on PyPI. Or if it could be merged into the qtreactor repo from which it originated. Though it seems the author on that repo is no longer responding to issues etc :( So maybe a new repo would be better.
1.0
Release qt5reactor separately - Would be great if the `qt5reactor.py` you folks have written could be broken out and put in a separate repo, packaged using setuptools and put on PyPI. Or if it could be merged into the qtreactor repo from which it originated. Though it seems the author on that repo is no longer responding to issues etc :( So maybe a new repo would be better.
process
release separately would be great if the py you folks have written could be broken out and put in a separate repo packaged using setuptools and put on pypi or if it could be merged into the qtreactor repo from which it originated though it seems the author on that repo is no longer responding to issues etc so maybe a new repo would be better
1
22,781
2,650,922,855
IssuesEvent
2015-03-16 06:47:34
grepper/tovid
https://api.github.com/repos/grepper/tovid
closed
wx to imagemagick font mapping misses too many matches
bug imported Priority-Medium wontfix
_From [da...@reallylongword.org](https://code.google.com/u/114094942197167775108/) on October 28, 2007 01:37:09_ The mapping between wx and ImageMagick fonts misses a lot. Getting the two font systems to play nice together is at its core educated guesses, but I think they could be improved. A lot of fonts are overlooked because the ImageMagick font names sometimes include a variant name at the end of the font, such as "Reguar", "Bold" or something along those lines, whereas the wx font names usually do not. I see two potential solutions: remove the variant name for the match, or match against the family name. I've created a modification to implement the second solution. If no match for a particular wx font name is found using the IM font name, a match is attempted using the IM style. I see two major problems with this patch. The first is the method of determining the font style. The output from "convert -list type" is meant to be human rather than machine readable; the font name is easy to match because it does not contain spaces, but the font family does. I use the header line to determine where the column splits are, and, since this line is in English, matching on words like "Family" and "Style" seems unreliable. The version of ImageMagick I'm using, 6.3.2.9, is not internationalized, so the words are going to be consistent regardless of locale, but this may not be true for later versions. Also, the order and names of the columns after "Name" might not be consistent across versions. I'm assuming that the columns always come in the order of "Name", "Version", "Style" and whatever else. Secondly, since names can be assumed to be unique but families cannot, this broader means of matching may create mappings from one wx font to more than one IM font. I've tried to mitigate this by eliminating duplicate IM font name values from the final mapping, but the means is somewhat capricious. I don't think this is too terrible, though, since the primary match is still based on name, so any font remaining from the duplicate elimination would at least be one variant of a font not otherwise available. I've attached a patch to libtovid/gui/configs.py against tovid-0.31. **Attachment:** [tovid-fonts.patch](http://code.google.com/p/tovid/issues/detail?id=19) _Original issue: http://code.google.com/p/tovid/issues/detail?id=19_
1.0
wx to imagemagick font mapping misses too many matches - _From [da...@reallylongword.org](https://code.google.com/u/114094942197167775108/) on October 28, 2007 01:37:09_ The mapping between wx and ImageMagick fonts misses a lot. Getting the two font systems to play nice together is at its core educated guesses, but I think they could be improved. A lot of fonts are overlooked because the ImageMagick font names sometimes include a variant name at the end of the font, such as "Reguar", "Bold" or something along those lines, whereas the wx font names usually do not. I see two potential solutions: remove the variant name for the match, or match against the family name. I've created a modification to implement the second solution. If no match for a particular wx font name is found using the IM font name, a match is attempted using the IM style. I see two major problems with this patch. The first is the method of determining the font style. The output from "convert -list type" is meant to be human rather than machine readable; the font name is easy to match because it does not contain spaces, but the font family does. I use the header line to determine where the column splits are, and, since this line is in English, matching on words like "Family" and "Style" seems unreliable. The version of ImageMagick I'm using, 6.3.2.9, is not internationalized, so the words are going to be consistent regardless of locale, but this may not be true for later versions. Also, the order and names of the columns after "Name" might not be consistent across versions. I'm assuming that the columns always come in the order of "Name", "Version", "Style" and whatever else. Secondly, since names can be assumed to be unique but families cannot, this broader means of matching may create mappings from one wx font to more than one IM font. I've tried to mitigate this by eliminating duplicate IM font name values from the final mapping, but the means is somewhat capricious. I don't think this is too terrible, though, since the primary match is still based on name, so any font remaining from the duplicate elimination would at least be one variant of a font not otherwise available. I've attached a patch to libtovid/gui/configs.py against tovid-0.31. **Attachment:** [tovid-fonts.patch](http://code.google.com/p/tovid/issues/detail?id=19) _Original issue: http://code.google.com/p/tovid/issues/detail?id=19_
non_process
wx to imagemagick font mapping misses too many matches from on october the mapping between wx and imagemagick fonts misses a lot getting the two font systems to play nice together is at its core educated guesses but i think they could be improved a lot of fonts are overlooked because the imagemagick font names sometimes include a variant name at the end of the font such as reguar bold or something along those lines whereas the wx font names usually do not i see two potential solutions remove the variant name for the match or match against the family name i ve created a modification to implement the second solution if no match for a particular wx font name is found using the im font name a match is attempted using the im style i see two major problems with this patch the first is the method of determining the font style the output from convert list type is meant to be human rather than machine readable the font name is easy to match because it does not contain spaces but the font family does i use the header line to determine where the column splits are and since this line is in english matching on words like family and style seems unreliable the version of imagemagick i m using is not internationalized so the words are going to be consistent regardless of locale but this may not be true for later versions also the order and names of the columns after name might not be consistent across versions i m assuming that the columns always come in the order of name version style and whatever else secondly since names can be assumed to be unique but families cannot this broader means of matching may create mappings from one wx font to more than one im font i ve tried to mitigate this by eliminating duplicate im font name values from the final mapping but the means is somewhat capricious i don t think this is too terrible though since the primary match is still based on name so any font remaining from the duplicate elimination would at least be one variant of a font not otherwise available i ve attached a patch to libtovid gui configs py against tovid attachment original issue
0
16,419
21,213,131,799
IssuesEvent
2022-04-11 03:00:35
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
gdal:warpreproject unexpected type handling
Feedback stale Processing Bug
### What is the bug or the crash? passing np.int32 parameters (instead of int) does not throw an error or recognize the input for gdal:warpreproject ### Steps to reproduce the issue ``` ins_d = { #other parameters 'TARGET_RESOLUTION': np.int32(100)} processing.run('gdal:warpreproject', ins_d) ``` does not throw an error and the gdalwarp command does not recognize the input `gdalwarp -of GTiff ...` while ``` ins_d = { #other parameters 'TARGET_RESOLUTION': int(100)} processing.run('gdal:warpreproject', ins_d) ``` prepares the gdal command as expected: `gdalwarp -tr 100.0 100.0 -of GTiff ...` ### Versions QGIS version 3.16.14-Hannover QGIS code revision 42791986 Compiled against Qt 5.15.2 Running against Qt 5.15.2 Compiled against GDAL/OGR 3.4.0 Running against GDAL/OGR 3.4.0 Compiled against GEOS 3.10.0-CAPI-1.16.0 Running against GEOS 3.10.0-CAPI-1.16.0 Compiled against SQLite 3.35.2 Running against SQLite 3.35.2 PostgreSQL Client Version 13.0 SpatiaLite Version 5.0.1 QWT Version 6.1.3 QScintilla2 Version 2.11.5 Compiled against PROJ 8.2.0 Running against PROJ Rel. 8.2.0, November 1st, 2021 OS Version Windows 10 Version 2009 Active python plugins canflood; firstaid; latlontools; MapsPrinter; menu_from_project; PluginLoadTimes; processing_taudem; processing_wbt; profiletool; db_manager; processing ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context This issue is especially nasty as visually inspecting the inputs does not show the type change. My expectation would be that the processing handler at least throws an error. maybe this should be a feature request? Either way, I hope this issue is helpful for others.
1.0
gdal:warpreproject unexpected type handling - ### What is the bug or the crash? passing np.int32 parameters (instead of int) does not throw an error or recognize the input for gdal:warpreproject ### Steps to reproduce the issue ``` ins_d = { #other parameters 'TARGET_RESOLUTION': np.int32(100)} processing.run('gdal:warpreproject', ins_d) ``` does not throw an error and the gdalwarp command does not recognize the input `gdalwarp -of GTiff ...` while ``` ins_d = { #other parameters 'TARGET_RESOLUTION': int(100)} processing.run('gdal:warpreproject', ins_d) ``` prepares the gdal command as expected: `gdalwarp -tr 100.0 100.0 -of GTiff ...` ### Versions QGIS version 3.16.14-Hannover QGIS code revision 42791986 Compiled against Qt 5.15.2 Running against Qt 5.15.2 Compiled against GDAL/OGR 3.4.0 Running against GDAL/OGR 3.4.0 Compiled against GEOS 3.10.0-CAPI-1.16.0 Running against GEOS 3.10.0-CAPI-1.16.0 Compiled against SQLite 3.35.2 Running against SQLite 3.35.2 PostgreSQL Client Version 13.0 SpatiaLite Version 5.0.1 QWT Version 6.1.3 QScintilla2 Version 2.11.5 Compiled against PROJ 8.2.0 Running against PROJ Rel. 8.2.0, November 1st, 2021 OS Version Windows 10 Version 2009 Active python plugins canflood; firstaid; latlontools; MapsPrinter; menu_from_project; PluginLoadTimes; processing_taudem; processing_wbt; profiletool; db_manager; processing ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context This issue is especially nasty as visually inspecting the inputs does not show the type change. My expectation would be that the processing handler at least throws an error. maybe this should be a feature request? Either way, I hope this issue is helpful for others.
process
gdal warpreproject unexpected type handling what is the bug or the crash passing np parameters instead of int does not throw an error or recognize the input for gdal warpreproject steps to reproduce the issue ins d other parameters target resolution np processing run gdal warpreproject ins d does not throw an error and the gdalwarp command does not recognize the input gdalwarp of gtiff while ins d other parameters target resolution int processing run gdal warpreproject ins d prepares the gdal command as expected gdalwarp tr of gtiff versions qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel november os version windows version active python plugins canflood firstaid latlontools mapsprinter menu from project pluginloadtimes processing taudem processing wbt profiletool db manager processing supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context this issue is especially nasty as visually inspecting the inputs does not show the type change my expectation would be that the processing handler at least throws an error maybe this should be a feature request either way i hope this issue is helpful for others
1
7,553
10,675,602,158
IssuesEvent
2019-10-21 12:04:50
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Process BasePriority test failure on WSL
area-System.Diagnostics.Process bug disabled-test os-windows-wsl
1 Process test consistently fails on WSL on the latest RS4 Insiders Preview build: ```` <test name="System.Diagnostics.Tests.ProcessTests.TestBasePriorityOnUnix" type="System.Diagnostics.Tests.ProcessTests" method="TestBasePriorityOnUnix" time="0.062201" result="Fail"> <traits> <trait name="category" value="requireselevation" /> </traits> <failure exception-type="Xunit.Sdk.EqualException"> <message><![CDATA[Assert.Equal() Failure\nExpected: 19\nActual: -19]]></message> <stack-trace><![CDATA[ at System.Diagnostics.Tests.ProcessTests.SetAndCheckBasePriority(ProcessPriorityClass exPriorityClass, Int32 priority) in /home/jfree/dd/corefx/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 45 at System.Diagnostics.Tests.ProcessTests.TestBasePriorityOnUnix() in /home/jfree/dd/corefx/src/System.Diagnostics.Process/tests/ProcessTests.Unix.cs:line 241]]></stack-trace> </failure> </test> ````
1.0
Process BasePriority test failure on WSL - 1 Process test consistently fails on WSL on the latest RS4 Insiders Preview build: ```` <test name="System.Diagnostics.Tests.ProcessTests.TestBasePriorityOnUnix" type="System.Diagnostics.Tests.ProcessTests" method="TestBasePriorityOnUnix" time="0.062201" result="Fail"> <traits> <trait name="category" value="requireselevation" /> </traits> <failure exception-type="Xunit.Sdk.EqualException"> <message><![CDATA[Assert.Equal() Failure\nExpected: 19\nActual: -19]]></message> <stack-trace><![CDATA[ at System.Diagnostics.Tests.ProcessTests.SetAndCheckBasePriority(ProcessPriorityClass exPriorityClass, Int32 priority) in /home/jfree/dd/corefx/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 45 at System.Diagnostics.Tests.ProcessTests.TestBasePriorityOnUnix() in /home/jfree/dd/corefx/src/System.Diagnostics.Process/tests/ProcessTests.Unix.cs:line 241]]></stack-trace> </failure> </test> ````
process
process basepriority test failure on wsl process test consistently fails on wsl on the latest insiders preview build cdata at system diagnostics tests processtests setandcheckbasepriority processpriorityclass expriorityclass priority in home jfree dd corefx src system diagnostics process tests processtests cs line at system diagnostics tests processtests testbasepriorityonunix in home jfree dd corefx src system diagnostics process tests processtests unix cs line
1
350,226
31,864,290,028
IssuesEvent
2023-09-15 13:09:02
agda/agda
https://api.github.com/repos/agda/agda
closed
`DuplicateExecutable` test randomly breaks other `Fail/` tests
infra: test suite
On latest `master` (283730b), running `make test` fails non-deterministically. The console pauses several times and displays diffs such as ```diff 1.actual index e83670021..a3a9e853d 100644 --- a/tmp/nix-shell.qnWrHw/AbstractModuleMacro524397-0.golden +++ b/tmp/nix-shell.qnWrHw/AbstractModuleMacro524397-1.actual @@ -1,5 +1,3 @@ -AbstractModuleMacro.agda:7,1-8,19 -warning: -W[no]UselessAbstract -Using abstract here has no effect. Abstract applies to only -definitions like data definitions, record type definitions and -function clauses. +Duplicate entries for executable 'echo' in DuplicateExecutable/executables: + 2: /bin/echo + 3: /bin/echo ``` or ```diff diff --git a/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-2.golden b/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-3.actual index dc1787cd6..a3a9e853d 100644 --- a/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-2.golden +++ b/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-3.actual @@ -1,3 +1,3 @@ -BadForcedIrrelevance.agda:11,16-17 -Variable n is declared erased, so it cannot be used here -when checking that the expression n has type Nat +Duplicate entries for executable 'echo' in DuplicateExecutable/executables: + 2: /bin/echo + 3: /bin/echo (END) ``` the incoming fragment is always for the `DuplicateExecutable` test. The other test varies, but is always from the `Fail/` suite. When running `make test -j1`, exactly 1 other test is affected, but without the `-j1` flag the number of affected tests varies. I am using GNU Make 4.4.1, cabal 3.8.1.0, and ghc 9.2.7, all provided by `nix develop` in this repo's root. OS is Ubuntu 22.04.3 LTS x86_64
1.0
`DuplicateExecutable` test randomly breaks other `Fail/` tests - On latest `master` (283730b), running `make test` fails non-deterministically. The console pauses several times and displays diffs such as ```diff 1.actual index e83670021..a3a9e853d 100644 --- a/tmp/nix-shell.qnWrHw/AbstractModuleMacro524397-0.golden +++ b/tmp/nix-shell.qnWrHw/AbstractModuleMacro524397-1.actual @@ -1,5 +1,3 @@ -AbstractModuleMacro.agda:7,1-8,19 -warning: -W[no]UselessAbstract -Using abstract here has no effect. Abstract applies to only -definitions like data definitions, record type definitions and -function clauses. +Duplicate entries for executable 'echo' in DuplicateExecutable/executables: + 2: /bin/echo + 3: /bin/echo ``` or ```diff diff --git a/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-2.golden b/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-3.actual index dc1787cd6..a3a9e853d 100644 --- a/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-2.golden +++ b/tmp/nix-shell.qnWrHw/BadForcedIrrelevance524397-3.actual @@ -1,3 +1,3 @@ -BadForcedIrrelevance.agda:11,16-17 -Variable n is declared erased, so it cannot be used here -when checking that the expression n has type Nat +Duplicate entries for executable 'echo' in DuplicateExecutable/executables: + 2: /bin/echo + 3: /bin/echo (END) ``` the incoming fragment is always for the `DuplicateExecutable` test. The other test varies, but is always from the `Fail/` suite. When running `make test -j1`, exactly 1 other test is affected, but without the `-j1` flag the number of affected tests varies. I am using GNU Make 4.4.1, cabal 3.8.1.0, and ghc 9.2.7, all provided by `nix develop` in this repo's root. OS is Ubuntu 22.04.3 LTS x86_64
non_process
duplicateexecutable test randomly breaks other fail tests on latest master running make test fails non deterministically the console pauses several times and displays diffs such as diff actual index a tmp nix shell qnwrhw golden b tmp nix shell qnwrhw actual abstractmodulemacro agda warning w uselessabstract using abstract here has no effect abstract applies to only definitions like data definitions record type definitions and function clauses duplicate entries for executable echo in duplicateexecutable executables bin echo bin echo or diff diff git a tmp nix shell qnwrhw golden b tmp nix shell qnwrhw actual index a tmp nix shell qnwrhw golden b tmp nix shell qnwrhw actual badforcedirrelevance agda variable n is declared erased so it cannot be used here when checking that the expression n has type nat duplicate entries for executable echo in duplicateexecutable executables bin echo bin echo end the incoming fragment is always for the duplicateexecutable test the other test varies but is always from the fail suite when running make test exactly other test is affected but without the flag the number of affected tests varies i am using gnu make cabal and ghc all provided by nix develop in this repo s root os is ubuntu lts
0
17,315
23,138,122,761
IssuesEvent
2022-07-28 15:51:01
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Add Hydrogen to "Other Fuels" for carbon
Process Heating important CO2 Emissions
Add Hydrogen to teh list of "Other Fuels" for carbon / CO2 calcs. It has a value of 0 CO2/MMBTU, but the work around for using it is a bit of a pain, so would be better if it had an entry.
1.0
Add Hydrogen to "Other Fuels" for carbon - Add Hydrogen to teh list of "Other Fuels" for carbon / CO2 calcs. It has a value of 0 CO2/MMBTU, but the work around for using it is a bit of a pain, so would be better if it had an entry.
process
add hydrogen to other fuels for carbon add hydrogen to teh list of other fuels for carbon calcs it has a value of mmbtu but the work around for using it is a bit of a pain so would be better if it had an entry
1
21,721
30,224,170,858
IssuesEvent
2023-07-05 22:09:40
LLNL/axom
https://api.github.com/repos/LLNL/axom
closed
Update automatic formatting to ensure one-liners are braced
Software process
We recently noticed that our automatic formatting (via `clang-format`) does not enforce braces around single-line control statements (e.g. `if`, `for`, ...). Since braces are required by our coding guide, our automatic formatting should enforce this. After a cursory glance, it looks like this feature was added to `clang-format` in clang@15 and we're currently using clang@10. The latest version of clang on LLNL's LC clusters is clang@14. It looks like we can use `clang-tidy` as a stopgap. See: * Coding guideline: > 9.1 Curly braces should be used in all conditionals, loops, etc. even when the content inside the braces is a “one-liner”. https://axom.readthedocs.io/en/develop/docs/sphinx/coding_guide/sec09_format.html * https://stackoverflow.com/questions/26111162/can-clang-format-add-braces-to-single-line-if-statements-etc
1.0
Update automatic formatting to ensure one-liners are braced - We recently noticed that our automatic formatting (via `clang-format`) does not enforce braces around single-line control statements (e.g. `if`, `for`, ...). Since braces are required by our coding guide, our automatic formatting should enforce this. After a cursory glance, it looks like this feature was added to `clang-format` in clang@15 and we're currently using clang@10. The latest version of clang on LLNL's LC clusters is clang@14. It looks like we can use `clang-tidy` as a stopgap. See: * Coding guideline: > 9.1 Curly braces should be used in all conditionals, loops, etc. even when the content inside the braces is a “one-liner”. https://axom.readthedocs.io/en/develop/docs/sphinx/coding_guide/sec09_format.html * https://stackoverflow.com/questions/26111162/can-clang-format-add-braces-to-single-line-if-statements-etc
process
update automatic formatting to ensure one liners are braced we recently noticed that our automatic formatting via clang format does not enforce braces around single line control statements e g if for since braces are required by our coding guide our automatic formatting should enforce this after a cursory glance it looks like this feature was added to clang format in clang and we re currently using clang the latest version of clang on llnl s lc clusters is clang it looks like we can use clang tidy as a stopgap see coding guideline curly braces should be used in all conditionals loops etc even when the content inside the braces is a “one liner”
1
20,918
27,755,555,559
IssuesEvent
2023-03-16 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 16 Mar 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Co-Occurrence Matters: Learning Action Relation for Temporal Action Localization - **Authors:** Congqi Cao, Yizhe Wang, Yue Lu, Xin Zhang, Yanning Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08463 - **Pdf link:** https://arxiv.org/pdf/2303.08463 - **Abstract** Temporal action localization (TAL) is a prevailing task due to its great application potential. Existing works in this field mainly suffer from two weaknesses: (1) They often neglect the multi-label case and only focus on temporal modeling. (2) They ignore the semantic information in class labels and only use the visual information. To solve these problems, we propose a novel Co-Occurrence Relation Module (CORM) that explicitly models the co-occurrence relationship between actions. Besides the visual information, it further utilizes the semantic embeddings of class labels to model the co-occurrence relationship. The CORM works in a plug-and-play manner and can be easily incorporated with the existing sequence models. By considering both visual and semantic co-occurrence, our method achieves high multi-label relationship modeling capacity. Meanwhile, existing datasets in TAL always focus on low-semantic atomic actions. Thus we construct a challenging multi-label dataset UCF-Crime-TAL that focuses on high-semantic actions by annotating the UCF-Crime dataset at frame level and considering the semantic overlap of different events. Extensive experiments on two commonly used TAL datasets, \textit{i.e.}, MultiTHUMOS and TSU, and our newly proposed UCF-Crime-TAL demenstrate the effectiveness of the proposed CORM, which achieves state-of-the-art performance on these datasets. ### Mutual Information-Based Temporal Difference Learning for Human Pose Estimation in Video - **Authors:** Runyang Feng, Yixing Gao, Xueqing Ma, Tze Ho Elden Tse, Hyung Jin Chang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08475 - **Pdf link:** https://arxiv.org/pdf/2303.08475 - **Abstract** Temporal modeling is crucial for multi-frame human pose estimation. Most existing methods directly employ optical flow or deformable convolution to predict full-spectrum motion fields, which might incur numerous irrelevant cues, such as a nearby person or background. Without further efforts to excavate meaningful motion priors, their results are suboptimal, especially in complicated spatiotemporal interactions. On the other hand, the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited. In this paper, we present a novel multi-frame human pose estimation framework, which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement. To be specific, we design a multi-stage Temporal Difference Encoder that performs incremental cascaded learning conditioned on multi-stage feature difference sequences to derive informative motion representation. We further propose a Representation Disentanglement module from the mutual information perspective, which can grasp discriminative task-relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information. These place us to rank No.1 in the Crowd Pose Estimation in Complex Events Challenge on benchmark dataset HiEve, and achieve state-of-the-art performance on three benchmarks PoseTrack2017, PoseTrack2018, and PoseTrack21. ## Keyword: event camera ### Improving Fast Auto-Focus with Event Polarity - **Authors:** Yuhan Bao, Lei Sun, Yuqin Ma, Diyang Gu, Kaiwei Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Optics (physics.optics) - **Arxiv link:** https://arxiv.org/abs/2303.08611 - **Pdf link:** https://arxiv.org/pdf/2303.08611 - **Abstract** Fast and accurate auto-focus in adverse conditions remains an arduous task. The paper presents a polarity-based event camera auto-focus algorithm featuring high-speed, precise auto-focus in dark, dynamic scenes that conventional frame-based cameras cannot match. Specifically, the symmetrical relationship between the event polarities in focusing is investigated, and the event-based focus evaluation function is proposed based on the principles of the event cameras and the imaging model in the focusing process. Comprehensive experiments on the public EAD dataset show the robustness of the model. Furthermore, precise focus with less than one depth of focus is achieved within 0.004 seconds on our self-built high-speed focusing platform. The dataset and code will be made publicly available. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Deep Visual Forced Alignment: Learning to Align Transcription with Talking Face Video - **Authors:** Minsu Kim, Chae Won Kim, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2303.08670 - **Pdf link:** https://arxiv.org/pdf/2303.08670 - **Abstract** Forced alignment refers to a technology that time-aligns a given transcription with a corresponding speech. However, as the forced alignment technologies have developed using speech audio, they might fail in alignment when the input speech audio is noise-corrupted or is not accessible. We focus on that there is another component that the speech can be inferred from, the speech video (i.e., talking face video). Since the drawbacks of audio-based forced alignment can be complemented using the visual information when the audio signal is under poor condition, we try to develop a novel video-based forced alignment method. However, different from audio forced alignment, it is challenging to develop a reliable visual forced alignment technology for the following two reasons: 1) Visual Speech Recognition (VSR) has a much lower performance compared to audio-based Automatic Speech Recognition (ASR), and 2) the translation from text to video is not reliable, so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment. In order to alleviate these challenges, in this paper, we propose a new method that is appropriate for visual forced alignment, namely Deep Visual Forced Alignment (DVFA). The proposed DVFA can align the input transcription (i.e., sentence) with the talking face video without accessing the speech audio. Moreover, by augmenting the alignment task with anomaly case detection, DVFA can detect mismatches between the input transcription and the input video while performing the alignment. Therefore, we can robustly align the text with the talking face video even if there exist error words in the text. Through extensive experiments, we show the effectiveness of the proposed DVFA not only in the alignment task but also in interpreting the outputs of VSR models. ## Keyword: ISP ### Rethinking Optical Flow from Geometric Matching Consistent Perspective - **Authors:** Qiaole Dong, Chenjie Cao, Yanwei Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08384 - **Pdf link:** https://arxiv.org/pdf/2303.08384 - **Abstract** Optical flow estimation is a challenging problem remaining unsolved. Recent deep learning based optical flow models have achieved considerable success. However, these models often train networks from the scratch on standard optical flow data, which restricts their ability to robustly and geometrically match image features. In this paper, we propose a rethinking to previous optical flow estimation. We particularly leverage Geometric Image Matching (GIM) as a pre-training task for the optical flow estimation (MatchFlow) with better feature representations, as GIM shares some common challenges as optical flow estimation, and with massive labeled real-world data. Thus, matching static scenes helps to learn more fundamental feature correlations of objects and scenes with consistent displacements. Specifically, the proposed MatchFlow model employs a QuadTree attention-based network pre-trained on MegaDepth to extract coarse features for further flow regression. Extensive experiments show that our model has great cross-dataset generalization. Our method achieves 11.5% and 10.1% error reduction from GMA on Sintel clean pass and KITTI test set. At the time of anonymous submission, our MatchFlow(G) enjoys state-of-the-art performance on Sintel clean and final pass compared to published approaches with comparable computation and memory footprint. Codes and models will be released in https://github.com/DQiaole/MatchFlow. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Mutual Information-Based Temporal Difference Learning for Human Pose Estimation in Video - **Authors:** Runyang Feng, Yixing Gao, Xueqing Ma, Tze Ho Elden Tse, Hyung Jin Chang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08475 - **Pdf link:** https://arxiv.org/pdf/2303.08475 - **Abstract** Temporal modeling is crucial for multi-frame human pose estimation. Most existing methods directly employ optical flow or deformable convolution to predict full-spectrum motion fields, which might incur numerous irrelevant cues, such as a nearby person or background. Without further efforts to excavate meaningful motion priors, their results are suboptimal, especially in complicated spatiotemporal interactions. On the other hand, the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited. In this paper, we present a novel multi-frame human pose estimation framework, which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement. To be specific, we design a multi-stage Temporal Difference Encoder that performs incremental cascaded learning conditioned on multi-stage feature difference sequences to derive informative motion representation. We further propose a Representation Disentanglement module from the mutual information perspective, which can grasp discriminative task-relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information. These place us to rank No.1 in the Crowd Pose Estimation in Complex Events Challenge on benchmark dataset HiEve, and achieve state-of-the-art performance on three benchmarks PoseTrack2017, PoseTrack2018, and PoseTrack21. ### Deep Visual Forced Alignment: Learning to Align Transcription with Talking Face Video - **Authors:** Minsu Kim, Chae Won Kim, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2303.08670 - **Pdf link:** https://arxiv.org/pdf/2303.08670 - **Abstract** Forced alignment refers to a technology that time-aligns a given transcription with a corresponding speech. However, as the forced alignment technologies have developed using speech audio, they might fail in alignment when the input speech audio is noise-corrupted or is not accessible. We focus on that there is another component that the speech can be inferred from, the speech video (i.e., talking face video). Since the drawbacks of audio-based forced alignment can be complemented using the visual information when the audio signal is under poor condition, we try to develop a novel video-based forced alignment method. However, different from audio forced alignment, it is challenging to develop a reliable visual forced alignment technology for the following two reasons: 1) Visual Speech Recognition (VSR) has a much lower performance compared to audio-based Automatic Speech Recognition (ASR), and 2) the translation from text to video is not reliable, so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment. In order to alleviate these challenges, in this paper, we propose a new method that is appropriate for visual forced alignment, namely Deep Visual Forced Alignment (DVFA). The proposed DVFA can align the input transcription (i.e., sentence) with the talking face video without accessing the speech audio. Moreover, by augmenting the alignment task with anomaly case detection, DVFA can detect mismatches between the input transcription and the input video while performing the alignment. Therefore, we can robustly align the text with the talking face video even if there exist error words in the text. Through extensive experiments, we show the effectiveness of the proposed DVFA not only in the alignment task but also in interpreting the outputs of VSR models. ### RSFNet: A White-Box Image Retouching Approach using Region-Specific Color Filters - **Authors:** Wenqi Ouyang, Yi Dong, Peiran Ren, Xiaoyang Kang, Xin Xu, Xuansong Xie - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08682 - **Pdf link:** https://arxiv.org/pdf/2303.08682 - **Abstract** Retouching images is an essential aspect of enhancing the visual appeal of photos. Although users often share common aesthetic preferences, their retouching methods may vary based on their individual preferences. Therefore, there is a need for white-box approaches that produce satisfying results and enable users to conveniently edit their images simultaneously. Recent white-box retouching methods rely on cascaded global filters that provide image-level filter arguments but cannot perform fine-grained retouching. In contrast, colorists typically use a divide-and-conquer approach, performing a series of region-specific fine-grained enhancements when using traditional tools like Davinci Resolve. We draw on this insight to develop a white-box framework for photo retouching using parallel region-specific filters, called RSFNet. Our model generates filter arguments (e.g., saturation, contrast, hue) and attention maps of regions for each filter simultaneously. Instead of cascading filters, RSFNet employs linear summations of filters, allowing for a more diverse range of filter classes that can be trained more easily. Our experiments demonstrate that RSFNet achieves state-of-the-art results, offering satisfying aesthetic appeal and greater user convenience for editable white-box retouching. ### Background Matters: Enhancing Out-of-distribution Detection with Domain Features - **Authors:** Choubo Ding, Guansong Pang, Chunhua Shen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08727 - **Pdf link:** https://arxiv.org/pdf/2303.08727 - **Abstract** Detecting out-of-distribution (OOD) inputs is a principal task for ensuring the safety of deploying deep-neural-network classifiers in open-world scenarios. OOD samples can be drawn from arbitrary distributions and exhibit deviations from in-distribution (ID) data in various dimensions, such as foreground semantic features (e.g., vehicle images vs. ID samples in fruit classification) and background domain features (e.g., textural images vs. ID samples in object recognition). Existing methods focus on detecting OOD samples based on the semantic features, while neglecting the other dimensions such as the domain features. This paper considers the importance of the domain features in OOD detection and proposes to leverage them to enhance the semantic-feature-based OOD detection methods. To this end, we propose a novel generic framework that can learn the domain features from the ID training samples by a dense prediction approach, with which different existing semantic-feature-based OOD detection methods can be seamlessly combined to jointly learn the in-distribution features from both the semantic and domain dimensions. Extensive experiments show that our approach 1) can substantially enhance the performance of four different state-of-the-art (SotA) OOD detection methods on multiple widely-used OOD datasets with diverse domain features, and 2) achieves new SotA performance on these benchmarks. ### Query-guided Attention in Vision Transformers for Localizing Objects Using a Single Sketch - **Authors:** Aditay Tripathi, Anand Mishra, Anirban Chakraborty - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08784 - **Pdf link:** https://arxiv.org/pdf/2303.08784 - **Abstract** In this work, we investigate the problem of sketch-based object localization on natural images, where given a crude hand-drawn sketch of an object, the goal is to localize all the instances of the same object on the target image. This problem proves difficult due to the abstract nature of hand-drawn sketches, variations in the style and quality of sketches, and the large domain gap existing between the sketches and the natural images. To mitigate these challenges, existing works proposed attention-based frameworks to incorporate query information into the image features. However, in these works, the query features are incorporated after the image features have already been independently learned, leading to inadequate alignment. In contrast, we propose a sketch-guided vision transformer encoder that uses cross-attention after each block of the transformer-based image encoder to learn query-conditioned image features leading to stronger alignment with the query sketch. Further, at the output of the decoder, the object and the sketch features are refined to bring the representation of relevant objects closer to the sketch query and thereby improve the localization. The proposed model also generalizes to the object categories not seen during training, as the target image features learned by our method are query-aware. Our localization framework can also utilize multiple sketch queries via a trainable novel sketch fusion strategy. The model is evaluated on the images from the public object detection benchmark, namely MS-COCO, using the sketch queries from QuickDraw! and Sketchy datasets. Compared with existing localization methods, the proposed approach gives a $6.6\%$ and $8.0\%$ improvement in mAP for seen objects using sketch queries from QuickDraw! and Sketchy datasets, respectively, and a $12.2\%$ improvement in AP@50 for large objects that are `unseen' during training. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 16 Mar 23 - ## Keyword: events ### Co-Occurrence Matters: Learning Action Relation for Temporal Action Localization - **Authors:** Congqi Cao, Yizhe Wang, Yue Lu, Xin Zhang, Yanning Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08463 - **Pdf link:** https://arxiv.org/pdf/2303.08463 - **Abstract** Temporal action localization (TAL) is a prevailing task due to its great application potential. Existing works in this field mainly suffer from two weaknesses: (1) They often neglect the multi-label case and only focus on temporal modeling. (2) They ignore the semantic information in class labels and only use the visual information. To solve these problems, we propose a novel Co-Occurrence Relation Module (CORM) that explicitly models the co-occurrence relationship between actions. Besides the visual information, it further utilizes the semantic embeddings of class labels to model the co-occurrence relationship. The CORM works in a plug-and-play manner and can be easily incorporated with the existing sequence models. By considering both visual and semantic co-occurrence, our method achieves high multi-label relationship modeling capacity. Meanwhile, existing datasets in TAL always focus on low-semantic atomic actions. Thus we construct a challenging multi-label dataset UCF-Crime-TAL that focuses on high-semantic actions by annotating the UCF-Crime dataset at frame level and considering the semantic overlap of different events. Extensive experiments on two commonly used TAL datasets, \textit{i.e.}, MultiTHUMOS and TSU, and our newly proposed UCF-Crime-TAL demenstrate the effectiveness of the proposed CORM, which achieves state-of-the-art performance on these datasets. ### Mutual Information-Based Temporal Difference Learning for Human Pose Estimation in Video - **Authors:** Runyang Feng, Yixing Gao, Xueqing Ma, Tze Ho Elden Tse, Hyung Jin Chang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08475 - **Pdf link:** https://arxiv.org/pdf/2303.08475 - **Abstract** Temporal modeling is crucial for multi-frame human pose estimation. Most existing methods directly employ optical flow or deformable convolution to predict full-spectrum motion fields, which might incur numerous irrelevant cues, such as a nearby person or background. Without further efforts to excavate meaningful motion priors, their results are suboptimal, especially in complicated spatiotemporal interactions. On the other hand, the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited. In this paper, we present a novel multi-frame human pose estimation framework, which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement. To be specific, we design a multi-stage Temporal Difference Encoder that performs incremental cascaded learning conditioned on multi-stage feature difference sequences to derive informative motion representation. We further propose a Representation Disentanglement module from the mutual information perspective, which can grasp discriminative task-relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information. These place us to rank No.1 in the Crowd Pose Estimation in Complex Events Challenge on benchmark dataset HiEve, and achieve state-of-the-art performance on three benchmarks PoseTrack2017, PoseTrack2018, and PoseTrack21. ## Keyword: event camera ### Improving Fast Auto-Focus with Event Polarity - **Authors:** Yuhan Bao, Lei Sun, Yuqin Ma, Diyang Gu, Kaiwei Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Optics (physics.optics) - **Arxiv link:** https://arxiv.org/abs/2303.08611 - **Pdf link:** https://arxiv.org/pdf/2303.08611 - **Abstract** Fast and accurate auto-focus in adverse conditions remains an arduous task. The paper presents a polarity-based event camera auto-focus algorithm featuring high-speed, precise auto-focus in dark, dynamic scenes that conventional frame-based cameras cannot match. Specifically, the symmetrical relationship between the event polarities in focusing is investigated, and the event-based focus evaluation function is proposed based on the principles of the event cameras and the imaging model in the focusing process. Comprehensive experiments on the public EAD dataset show the robustness of the model. Furthermore, precise focus with less than one depth of focus is achieved within 0.004 seconds on our self-built high-speed focusing platform. The dataset and code will be made publicly available. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Deep Visual Forced Alignment: Learning to Align Transcription with Talking Face Video - **Authors:** Minsu Kim, Chae Won Kim, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2303.08670 - **Pdf link:** https://arxiv.org/pdf/2303.08670 - **Abstract** Forced alignment refers to a technology that time-aligns a given transcription with a corresponding speech. However, as the forced alignment technologies have developed using speech audio, they might fail in alignment when the input speech audio is noise-corrupted or is not accessible. We focus on that there is another component that the speech can be inferred from, the speech video (i.e., talking face video). Since the drawbacks of audio-based forced alignment can be complemented using the visual information when the audio signal is under poor condition, we try to develop a novel video-based forced alignment method. However, different from audio forced alignment, it is challenging to develop a reliable visual forced alignment technology for the following two reasons: 1) Visual Speech Recognition (VSR) has a much lower performance compared to audio-based Automatic Speech Recognition (ASR), and 2) the translation from text to video is not reliable, so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment. In order to alleviate these challenges, in this paper, we propose a new method that is appropriate for visual forced alignment, namely Deep Visual Forced Alignment (DVFA). The proposed DVFA can align the input transcription (i.e., sentence) with the talking face video without accessing the speech audio. Moreover, by augmenting the alignment task with anomaly case detection, DVFA can detect mismatches between the input transcription and the input video while performing the alignment. Therefore, we can robustly align the text with the talking face video even if there exist error words in the text. Through extensive experiments, we show the effectiveness of the proposed DVFA not only in the alignment task but also in interpreting the outputs of VSR models. ## Keyword: ISP ### Rethinking Optical Flow from Geometric Matching Consistent Perspective - **Authors:** Qiaole Dong, Chenjie Cao, Yanwei Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08384 - **Pdf link:** https://arxiv.org/pdf/2303.08384 - **Abstract** Optical flow estimation is a challenging problem remaining unsolved. Recent deep learning based optical flow models have achieved considerable success. However, these models often train networks from the scratch on standard optical flow data, which restricts their ability to robustly and geometrically match image features. In this paper, we propose a rethinking to previous optical flow estimation. We particularly leverage Geometric Image Matching (GIM) as a pre-training task for the optical flow estimation (MatchFlow) with better feature representations, as GIM shares some common challenges as optical flow estimation, and with massive labeled real-world data. Thus, matching static scenes helps to learn more fundamental feature correlations of objects and scenes with consistent displacements. Specifically, the proposed MatchFlow model employs a QuadTree attention-based network pre-trained on MegaDepth to extract coarse features for further flow regression. Extensive experiments show that our model has great cross-dataset generalization. Our method achieves 11.5% and 10.1% error reduction from GMA on Sintel clean pass and KITTI test set. At the time of anonymous submission, our MatchFlow(G) enjoys state-of-the-art performance on Sintel clean and final pass compared to published approaches with comparable computation and memory footprint. Codes and models will be released in https://github.com/DQiaole/MatchFlow. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Mutual Information-Based Temporal Difference Learning for Human Pose Estimation in Video - **Authors:** Runyang Feng, Yixing Gao, Xueqing Ma, Tze Ho Elden Tse, Hyung Jin Chang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08475 - **Pdf link:** https://arxiv.org/pdf/2303.08475 - **Abstract** Temporal modeling is crucial for multi-frame human pose estimation. Most existing methods directly employ optical flow or deformable convolution to predict full-spectrum motion fields, which might incur numerous irrelevant cues, such as a nearby person or background. Without further efforts to excavate meaningful motion priors, their results are suboptimal, especially in complicated spatiotemporal interactions. On the other hand, the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited. In this paper, we present a novel multi-frame human pose estimation framework, which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement. To be specific, we design a multi-stage Temporal Difference Encoder that performs incremental cascaded learning conditioned on multi-stage feature difference sequences to derive informative motion representation. We further propose a Representation Disentanglement module from the mutual information perspective, which can grasp discriminative task-relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information. These place us to rank No.1 in the Crowd Pose Estimation in Complex Events Challenge on benchmark dataset HiEve, and achieve state-of-the-art performance on three benchmarks PoseTrack2017, PoseTrack2018, and PoseTrack21. ### Deep Visual Forced Alignment: Learning to Align Transcription with Talking Face Video - **Authors:** Minsu Kim, Chae Won Kim, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2303.08670 - **Pdf link:** https://arxiv.org/pdf/2303.08670 - **Abstract** Forced alignment refers to a technology that time-aligns a given transcription with a corresponding speech. However, as the forced alignment technologies have developed using speech audio, they might fail in alignment when the input speech audio is noise-corrupted or is not accessible. We focus on that there is another component that the speech can be inferred from, the speech video (i.e., talking face video). Since the drawbacks of audio-based forced alignment can be complemented using the visual information when the audio signal is under poor condition, we try to develop a novel video-based forced alignment method. However, different from audio forced alignment, it is challenging to develop a reliable visual forced alignment technology for the following two reasons: 1) Visual Speech Recognition (VSR) has a much lower performance compared to audio-based Automatic Speech Recognition (ASR), and 2) the translation from text to video is not reliable, so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment. In order to alleviate these challenges, in this paper, we propose a new method that is appropriate for visual forced alignment, namely Deep Visual Forced Alignment (DVFA). The proposed DVFA can align the input transcription (i.e., sentence) with the talking face video without accessing the speech audio. Moreover, by augmenting the alignment task with anomaly case detection, DVFA can detect mismatches between the input transcription and the input video while performing the alignment. Therefore, we can robustly align the text with the talking face video even if there exist error words in the text. Through extensive experiments, we show the effectiveness of the proposed DVFA not only in the alignment task but also in interpreting the outputs of VSR models. ### RSFNet: A White-Box Image Retouching Approach using Region-Specific Color Filters - **Authors:** Wenqi Ouyang, Yi Dong, Peiran Ren, Xiaoyang Kang, Xin Xu, Xuansong Xie - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08682 - **Pdf link:** https://arxiv.org/pdf/2303.08682 - **Abstract** Retouching images is an essential aspect of enhancing the visual appeal of photos. Although users often share common aesthetic preferences, their retouching methods may vary based on their individual preferences. Therefore, there is a need for white-box approaches that produce satisfying results and enable users to conveniently edit their images simultaneously. Recent white-box retouching methods rely on cascaded global filters that provide image-level filter arguments but cannot perform fine-grained retouching. In contrast, colorists typically use a divide-and-conquer approach, performing a series of region-specific fine-grained enhancements when using traditional tools like Davinci Resolve. We draw on this insight to develop a white-box framework for photo retouching using parallel region-specific filters, called RSFNet. Our model generates filter arguments (e.g., saturation, contrast, hue) and attention maps of regions for each filter simultaneously. Instead of cascading filters, RSFNet employs linear summations of filters, allowing for a more diverse range of filter classes that can be trained more easily. Our experiments demonstrate that RSFNet achieves state-of-the-art results, offering satisfying aesthetic appeal and greater user convenience for editable white-box retouching. ### Background Matters: Enhancing Out-of-distribution Detection with Domain Features - **Authors:** Choubo Ding, Guansong Pang, Chunhua Shen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08727 - **Pdf link:** https://arxiv.org/pdf/2303.08727 - **Abstract** Detecting out-of-distribution (OOD) inputs is a principal task for ensuring the safety of deploying deep-neural-network classifiers in open-world scenarios. OOD samples can be drawn from arbitrary distributions and exhibit deviations from in-distribution (ID) data in various dimensions, such as foreground semantic features (e.g., vehicle images vs. ID samples in fruit classification) and background domain features (e.g., textural images vs. ID samples in object recognition). Existing methods focus on detecting OOD samples based on the semantic features, while neglecting the other dimensions such as the domain features. This paper considers the importance of the domain features in OOD detection and proposes to leverage them to enhance the semantic-feature-based OOD detection methods. To this end, we propose a novel generic framework that can learn the domain features from the ID training samples by a dense prediction approach, with which different existing semantic-feature-based OOD detection methods can be seamlessly combined to jointly learn the in-distribution features from both the semantic and domain dimensions. Extensive experiments show that our approach 1) can substantially enhance the performance of four different state-of-the-art (SotA) OOD detection methods on multiple widely-used OOD datasets with diverse domain features, and 2) achieves new SotA performance on these benchmarks. ### Query-guided Attention in Vision Transformers for Localizing Objects Using a Single Sketch - **Authors:** Aditay Tripathi, Anand Mishra, Anirban Chakraborty - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2303.08784 - **Pdf link:** https://arxiv.org/pdf/2303.08784 - **Abstract** In this work, we investigate the problem of sketch-based object localization on natural images, where given a crude hand-drawn sketch of an object, the goal is to localize all the instances of the same object on the target image. This problem proves difficult due to the abstract nature of hand-drawn sketches, variations in the style and quality of sketches, and the large domain gap existing between the sketches and the natural images. To mitigate these challenges, existing works proposed attention-based frameworks to incorporate query information into the image features. However, in these works, the query features are incorporated after the image features have already been independently learned, leading to inadequate alignment. In contrast, we propose a sketch-guided vision transformer encoder that uses cross-attention after each block of the transformer-based image encoder to learn query-conditioned image features leading to stronger alignment with the query sketch. Further, at the output of the decoder, the object and the sketch features are refined to bring the representation of relevant objects closer to the sketch query and thereby improve the localization. The proposed model also generalizes to the object categories not seen during training, as the target image features learned by our method are query-aware. Our localization framework can also utilize multiple sketch queries via a trainable novel sketch fusion strategy. The model is evaluated on the images from the public object detection benchmark, namely MS-COCO, using the sketch queries from QuickDraw! and Sketchy datasets. Compared with existing localization methods, the proposed approach gives a $6.6\%$ and $8.0\%$ improvement in mAP for seen objects using sketch queries from QuickDraw! and Sketchy datasets, respectively, and a $12.2\%$ improvement in AP@50 for large objects that are `unseen' during training. ## Keyword: raw image There is no result
process
new submissions for thu mar keyword events co occurrence matters learning action relation for temporal action localization authors congqi cao yizhe wang yue lu xin zhang yanning zhang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract temporal action localization tal is a prevailing task due to its great application potential existing works in this field mainly suffer from two weaknesses they often neglect the multi label case and only focus on temporal modeling they ignore the semantic information in class labels and only use the visual information to solve these problems we propose a novel co occurrence relation module corm that explicitly models the co occurrence relationship between actions besides the visual information it further utilizes the semantic embeddings of class labels to model the co occurrence relationship the corm works in a plug and play manner and can be easily incorporated with the existing sequence models by considering both visual and semantic co occurrence our method achieves high multi label relationship modeling capacity meanwhile existing datasets in tal always focus on low semantic atomic actions thus we construct a challenging multi label dataset ucf crime tal that focuses on high semantic actions by annotating the ucf crime dataset at frame level and considering the semantic overlap of different events extensive experiments on two commonly used tal datasets textit i e multithumos and tsu and our newly proposed ucf crime tal demenstrate the effectiveness of the proposed corm which achieves state of the art performance on these datasets mutual information based temporal difference learning for human pose estimation in video authors runyang feng yixing gao xueqing ma tze ho elden tse hyung jin chang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract temporal modeling is crucial for multi frame human pose estimation most existing methods directly employ optical flow or deformable convolution to predict full spectrum motion fields which might incur numerous irrelevant cues such as a nearby person or background without further efforts to excavate meaningful motion priors their results are suboptimal especially in complicated spatiotemporal interactions on the other hand the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited in this paper we present a novel multi frame human pose estimation framework which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement to be specific we design a multi stage temporal difference encoder that performs incremental cascaded learning conditioned on multi stage feature difference sequences to derive informative motion representation we further propose a representation disentanglement module from the mutual information perspective which can grasp discriminative task relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information these place us to rank no in the crowd pose estimation in complex events challenge on benchmark dataset hieve and achieve state of the art performance on three benchmarks and keyword event camera improving fast auto focus with event polarity authors yuhan bao lei sun yuqin ma diyang gu kaiwei wang subjects computer vision and pattern recognition cs cv optics physics optics arxiv link pdf link abstract fast and accurate auto focus in adverse conditions remains an arduous task the paper presents a polarity based event camera auto focus algorithm featuring high speed precise auto focus in dark dynamic scenes that conventional frame based cameras cannot match specifically the symmetrical relationship between the event polarities in focusing is investigated and the event based focus evaluation function is proposed based on the principles of the event cameras and the imaging model in the focusing process comprehensive experiments on the public ead dataset show the robustness of the model furthermore precise focus with less than one depth of focus is achieved within seconds on our self built high speed focusing platform the dataset and code will be made publicly available keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb deep visual forced alignment learning to align transcription with talking face video authors minsu kim chae won kim yong man ro subjects computer vision and pattern recognition cs cv artificial intelligence cs ai sound cs sd audio and speech processing eess as arxiv link pdf link abstract forced alignment refers to a technology that time aligns a given transcription with a corresponding speech however as the forced alignment technologies have developed using speech audio they might fail in alignment when the input speech audio is noise corrupted or is not accessible we focus on that there is another component that the speech can be inferred from the speech video i e talking face video since the drawbacks of audio based forced alignment can be complemented using the visual information when the audio signal is under poor condition we try to develop a novel video based forced alignment method however different from audio forced alignment it is challenging to develop a reliable visual forced alignment technology for the following two reasons visual speech recognition vsr has a much lower performance compared to audio based automatic speech recognition asr and the translation from text to video is not reliable so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment in order to alleviate these challenges in this paper we propose a new method that is appropriate for visual forced alignment namely deep visual forced alignment dvfa the proposed dvfa can align the input transcription i e sentence with the talking face video without accessing the speech audio moreover by augmenting the alignment task with anomaly case detection dvfa can detect mismatches between the input transcription and the input video while performing the alignment therefore we can robustly align the text with the talking face video even if there exist error words in the text through extensive experiments we show the effectiveness of the proposed dvfa not only in the alignment task but also in interpreting the outputs of vsr models keyword isp rethinking optical flow from geometric matching consistent perspective authors qiaole dong chenjie cao yanwei fu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract optical flow estimation is a challenging problem remaining unsolved recent deep learning based optical flow models have achieved considerable success however these models often train networks from the scratch on standard optical flow data which restricts their ability to robustly and geometrically match image features in this paper we propose a rethinking to previous optical flow estimation we particularly leverage geometric image matching gim as a pre training task for the optical flow estimation matchflow with better feature representations as gim shares some common challenges as optical flow estimation and with massive labeled real world data thus matching static scenes helps to learn more fundamental feature correlations of objects and scenes with consistent displacements specifically the proposed matchflow model employs a quadtree attention based network pre trained on megadepth to extract coarse features for further flow regression extensive experiments show that our model has great cross dataset generalization our method achieves and error reduction from gma on sintel clean pass and kitti test set at the time of anonymous submission our matchflow g enjoys state of the art performance on sintel clean and final pass compared to published approaches with comparable computation and memory footprint codes and models will be released in keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw mutual information based temporal difference learning for human pose estimation in video authors runyang feng yixing gao xueqing ma tze ho elden tse hyung jin chang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract temporal modeling is crucial for multi frame human pose estimation most existing methods directly employ optical flow or deformable convolution to predict full spectrum motion fields which might incur numerous irrelevant cues such as a nearby person or background without further efforts to excavate meaningful motion priors their results are suboptimal especially in complicated spatiotemporal interactions on the other hand the temporal difference has the ability to encode representative motion information which can potentially be valuable for pose estimation but has not been fully exploited in this paper we present a novel multi frame human pose estimation framework which employs temporal differences across frames to model dynamic contexts and engages mutual information objectively to facilitate useful motion information disentanglement to be specific we design a multi stage temporal difference encoder that performs incremental cascaded learning conditioned on multi stage feature difference sequences to derive informative motion representation we further propose a representation disentanglement module from the mutual information perspective which can grasp discriminative task relevant motion signals by explicitly defining useful and noisy constituents of the raw motion features and minimizing their mutual information these place us to rank no in the crowd pose estimation in complex events challenge on benchmark dataset hieve and achieve state of the art performance on three benchmarks and deep visual forced alignment learning to align transcription with talking face video authors minsu kim chae won kim yong man ro subjects computer vision and pattern recognition cs cv artificial intelligence cs ai sound cs sd audio and speech processing eess as arxiv link pdf link abstract forced alignment refers to a technology that time aligns a given transcription with a corresponding speech however as the forced alignment technologies have developed using speech audio they might fail in alignment when the input speech audio is noise corrupted or is not accessible we focus on that there is another component that the speech can be inferred from the speech video i e talking face video since the drawbacks of audio based forced alignment can be complemented using the visual information when the audio signal is under poor condition we try to develop a novel video based forced alignment method however different from audio forced alignment it is challenging to develop a reliable visual forced alignment technology for the following two reasons visual speech recognition vsr has a much lower performance compared to audio based automatic speech recognition asr and the translation from text to video is not reliable so the method typically used for building audio forced alignment cannot be utilized in developing visual forced alignment in order to alleviate these challenges in this paper we propose a new method that is appropriate for visual forced alignment namely deep visual forced alignment dvfa the proposed dvfa can align the input transcription i e sentence with the talking face video without accessing the speech audio moreover by augmenting the alignment task with anomaly case detection dvfa can detect mismatches between the input transcription and the input video while performing the alignment therefore we can robustly align the text with the talking face video even if there exist error words in the text through extensive experiments we show the effectiveness of the proposed dvfa not only in the alignment task but also in interpreting the outputs of vsr models rsfnet a white box image retouching approach using region specific color filters authors wenqi ouyang yi dong peiran ren xiaoyang kang xin xu xuansong xie subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract retouching images is an essential aspect of enhancing the visual appeal of photos although users often share common aesthetic preferences their retouching methods may vary based on their individual preferences therefore there is a need for white box approaches that produce satisfying results and enable users to conveniently edit their images simultaneously recent white box retouching methods rely on cascaded global filters that provide image level filter arguments but cannot perform fine grained retouching in contrast colorists typically use a divide and conquer approach performing a series of region specific fine grained enhancements when using traditional tools like davinci resolve we draw on this insight to develop a white box framework for photo retouching using parallel region specific filters called rsfnet our model generates filter arguments e g saturation contrast hue and attention maps of regions for each filter simultaneously instead of cascading filters rsfnet employs linear summations of filters allowing for a more diverse range of filter classes that can be trained more easily our experiments demonstrate that rsfnet achieves state of the art results offering satisfying aesthetic appeal and greater user convenience for editable white box retouching background matters enhancing out of distribution detection with domain features authors choubo ding guansong pang chunhua shen subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract detecting out of distribution ood inputs is a principal task for ensuring the safety of deploying deep neural network classifiers in open world scenarios ood samples can be drawn from arbitrary distributions and exhibit deviations from in distribution id data in various dimensions such as foreground semantic features e g vehicle images vs id samples in fruit classification and background domain features e g textural images vs id samples in object recognition existing methods focus on detecting ood samples based on the semantic features while neglecting the other dimensions such as the domain features this paper considers the importance of the domain features in ood detection and proposes to leverage them to enhance the semantic feature based ood detection methods to this end we propose a novel generic framework that can learn the domain features from the id training samples by a dense prediction approach with which different existing semantic feature based ood detection methods can be seamlessly combined to jointly learn the in distribution features from both the semantic and domain dimensions extensive experiments show that our approach can substantially enhance the performance of four different state of the art sota ood detection methods on multiple widely used ood datasets with diverse domain features and achieves new sota performance on these benchmarks query guided attention in vision transformers for localizing objects using a single sketch authors aditay tripathi anand mishra anirban chakraborty subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in this work we investigate the problem of sketch based object localization on natural images where given a crude hand drawn sketch of an object the goal is to localize all the instances of the same object on the target image this problem proves difficult due to the abstract nature of hand drawn sketches variations in the style and quality of sketches and the large domain gap existing between the sketches and the natural images to mitigate these challenges existing works proposed attention based frameworks to incorporate query information into the image features however in these works the query features are incorporated after the image features have already been independently learned leading to inadequate alignment in contrast we propose a sketch guided vision transformer encoder that uses cross attention after each block of the transformer based image encoder to learn query conditioned image features leading to stronger alignment with the query sketch further at the output of the decoder the object and the sketch features are refined to bring the representation of relevant objects closer to the sketch query and thereby improve the localization the proposed model also generalizes to the object categories not seen during training as the target image features learned by our method are query aware our localization framework can also utilize multiple sketch queries via a trainable novel sketch fusion strategy the model is evaluated on the images from the public object detection benchmark namely ms coco using the sketch queries from quickdraw and sketchy datasets compared with existing localization methods the proposed approach gives a and improvement in map for seen objects using sketch queries from quickdraw and sketchy datasets respectively and a improvement in ap for large objects that are unseen during training keyword raw image there is no result
1
204,099
15,893,915,982
IssuesEvent
2021-04-11 08:10:55
digitallyinduced/ihp
https://api.github.com/repos/digitallyinduced/ihp
closed
Docu Section on Special Haskell Operations
documentation
[From reddit](https://www.reddit.com/r/haskell/comments/lt2u0a/ann_ihp_v09_has_been_released/gp0otu1/?utm_source=reddit&utm_medium=web2x&context=3) > Is there anywhere that explains some of the "exotic" language features of ihp? As a still somewhat newbie haskeller there are a few things I found while reading the docs that I couldn't understand. For example: ```haskell set #companyId (get #companyId company) query @Post fill @["title","body"] ``` > What are these @/# characters doing? It's frustrating as a newbie to see these things and not have any easy way of figuring them out. I can't even use Google/hoogle to search those. > html ShowView { .. } was initially confusing as well but I figured that one out.
1.0
Docu Section on Special Haskell Operations - [From reddit](https://www.reddit.com/r/haskell/comments/lt2u0a/ann_ihp_v09_has_been_released/gp0otu1/?utm_source=reddit&utm_medium=web2x&context=3) > Is there anywhere that explains some of the "exotic" language features of ihp? As a still somewhat newbie haskeller there are a few things I found while reading the docs that I couldn't understand. For example: ```haskell set #companyId (get #companyId company) query @Post fill @["title","body"] ``` > What are these @/# characters doing? It's frustrating as a newbie to see these things and not have any easy way of figuring them out. I can't even use Google/hoogle to search those. > html ShowView { .. } was initially confusing as well but I figured that one out.
non_process
docu section on special haskell operations is there anywhere that explains some of the exotic language features of ihp as a still somewhat newbie haskeller there are a few things i found while reading the docs that i couldn t understand for example haskell set companyid get companyid company query post fill what are these characters doing it s frustrating as a newbie to see these things and not have any easy way of figuring them out i can t even use google hoogle to search those html showview was initially confusing as well but i figured that one out
0
108,158
11,581,709,609
IssuesEvent
2020-02-21 23:31:28
sabaronett/reboundx
https://api.github.com/repos/sabaronett/reboundx
closed
Duplicate Header Comment (modify_mass)
documentation
Discovered the header comment to `examples/modify_mass/problem.c` is identical/taken from `/modify_orbits`. https://github.com/sabaronett/reboundx/blob/a0cf02bd1811aeb41e50535372a3fa9576b393b7/examples/modify_mass/problem.c#L1-L9
1.0
Duplicate Header Comment (modify_mass) - Discovered the header comment to `examples/modify_mass/problem.c` is identical/taken from `/modify_orbits`. https://github.com/sabaronett/reboundx/blob/a0cf02bd1811aeb41e50535372a3fa9576b393b7/examples/modify_mass/problem.c#L1-L9
non_process
duplicate header comment modify mass discovered the header comment to examples modify mass problem c is identical taken from modify orbits
0
12,368
14,894,955,181
IssuesEvent
2021-01-21 08:24:19
Maximus5/ConEmu
https://api.github.com/repos/Maximus5/ConEmu
reopened
201124: Failed exec batch files when path contained space and @
processes
conemu: 201124 x86 far: 3.0.305709 x86 win 10x64 Не запускаются батники в папках с символами пробела и @. Пример: Cоздаём папку с символами пробела и @ `md "C:\1 @" ` Cоздаём в папке выше cmd-файл "a.cmd": ` echo echo OK>a.cmd ` Пытаемся его запустить: ``` C:\1 @\@>a.cmd 'C:\1' is not recognized as an internal or external command, operable program or batch file. ```
1.0
201124: Failed exec batch files when path contained space and @ - conemu: 201124 x86 far: 3.0.305709 x86 win 10x64 Не запускаются батники в папках с символами пробела и @. Пример: Cоздаём папку с символами пробела и @ `md "C:\1 @" ` Cоздаём в папке выше cmd-файл "a.cmd": ` echo echo OK>a.cmd ` Пытаемся его запустить: ``` C:\1 @\@>a.cmd 'C:\1' is not recognized as an internal or external command, operable program or batch file. ```
process
failed exec batch files when path contained space and conemu far win не запускаются батники в папках с символами пробела и пример cоздаём папку с символами пробела и md c cоздаём в папке выше cmd файл a cmd echo echo ok a cmd пытаемся его запустить c a cmd c is not recognized as an internal or external command operable program or batch file
1
20,755
10,917,909,976
IssuesEvent
2019-11-21 15:57:37
magento/pwa-studio
https://api.github.com/repos/magento/pwa-studio
closed
[feature]: Implement product image pre fetching.
Progress: PR created enhancement performance
**Is your feature request related to a problem? Please describe.** Related to #1673 **Describe the solution you'd like** As of today, we do not prefetch images on the products page. For instance if the user is looking at image number X, we can make the service worker request images X-1 and X+1 for a smoother and faster experience. This can be achieved by the react component as well. We can make the call about where to implement this logic. This might be a far fetch but we can prefetch the corresponding image of a different color. For instance if the user is viewing the green version of image X, prefetch the other versions of X as well. The reason I call this a far fetch is that we will be prefetching `number_of_colors * 3` every time (X-1, X, X+1). Again there will be incremental benefits to this but the initial investment is high due to too many prefetches. Ideally, this only needs to be implemented on devices connected to WiFi. **Describe alternatives you've considered** Implementing this logic in https://github.com/magento/pwa-studio/blob/develop/packages/venia-ui/lib/components/ProductImageCarousel/carousel.js **Please let us know what packages this feature is in regards to:** - [x] `venia-concept` - [x] `venia-ui` - [ ] `pwa-buildpack` - [x] `peregrine` - [ ] `pwa-devdocs` - [ ] `upward-js` - [ ] `upward-spec`
True
[feature]: Implement product image pre fetching. - **Is your feature request related to a problem? Please describe.** Related to #1673 **Describe the solution you'd like** As of today, we do not prefetch images on the products page. For instance if the user is looking at image number X, we can make the service worker request images X-1 and X+1 for a smoother and faster experience. This can be achieved by the react component as well. We can make the call about where to implement this logic. This might be a far fetch but we can prefetch the corresponding image of a different color. For instance if the user is viewing the green version of image X, prefetch the other versions of X as well. The reason I call this a far fetch is that we will be prefetching `number_of_colors * 3` every time (X-1, X, X+1). Again there will be incremental benefits to this but the initial investment is high due to too many prefetches. Ideally, this only needs to be implemented on devices connected to WiFi. **Describe alternatives you've considered** Implementing this logic in https://github.com/magento/pwa-studio/blob/develop/packages/venia-ui/lib/components/ProductImageCarousel/carousel.js **Please let us know what packages this feature is in regards to:** - [x] `venia-concept` - [x] `venia-ui` - [ ] `pwa-buildpack` - [x] `peregrine` - [ ] `pwa-devdocs` - [ ] `upward-js` - [ ] `upward-spec`
non_process
implement product image pre fetching is your feature request related to a problem please describe related to describe the solution you d like as of today we do not prefetch images on the products page for instance if the user is looking at image number x we can make the service worker request images x and x for a smoother and faster experience this can be achieved by the react component as well we can make the call about where to implement this logic this might be a far fetch but we can prefetch the corresponding image of a different color for instance if the user is viewing the green version of image x prefetch the other versions of x as well the reason i call this a far fetch is that we will be prefetching number of colors every time x x x again there will be incremental benefits to this but the initial investment is high due to too many prefetches ideally this only needs to be implemented on devices connected to wifi describe alternatives you ve considered implementing this logic in please let us know what packages this feature is in regards to venia concept venia ui pwa buildpack peregrine pwa devdocs upward js upward spec
0
69,546
14,996,903,782
IssuesEvent
2021-01-29 16:11:17
cockroachdb/docs
https://api.github.com/repos/cockroachdb/docs
opened
pgwire: make it possible for a SQL proxy to override the client addr
A-security-&-iam C-release-note
PR: https://github.com/cockroachdb/cockroach/pull/58381 When using a SQL proxy, in the default configuration CockroachDB only knows about the network address of the proxy. That *peer* address is then used for logging, [authentication](../v21.1/authentication.html) rules, etc. This is undesirable, as security logging and authentication rules need to operate on the actual (final) client address instead. CockroachDB can now be configured to solve this problem (conf mechanism detailed below). When so configured, a SQL proxy can inform the CockroachDB server of the real address of the client via a server status parameter called `crdb:remote_addr`. The value must be the IP address of the client, followed by a colon, followed by the port number, using the standard Go syntax (e.g., `11.22.33.44:5566` for IPv4, `[11:22::33]:4455` for IPv6). When provided, this value overrides the SQL proxy's address for logging and authentication purposes. In any case, the original peer address is also logged alongside the client address (overridden or not), via the new logging tag `peer`. Security considerations: - Enabling this feature allows the peer to spoof its address with regard to authentication and thus bypass authentication rules that would otherwise apply to its address, which can introduce a serious security vulnerability if the peer is not trusted. This is why this feature is not enabled by default, and must only be enabled when using a trusted SQL proxy. - This feature should only be used with SQL proxies which actively scrub a `crdb:remote_addr` parameter received by a remote client, and replaces it by its own. If the proxy mistakenly forwards the parameter as provided by the client, it opens the door to the aforementioned security vulnerability. - Care must be taken in HBA rules: TLS client cert validation, if requested by a rule, is still performed using the certificate presented by the proxy, not that presented by the client. This means that this new feature is not sufficient to forward TLS client cert authentication through a proxy. (If TLS client cert authentication is required, it must be performed by the proxy directly.) - Care must be taken in HBA rules: the `protocol` field (first column) continues to apply to the connection type between CockroachDB and the proxy, not between the proxy and the client. Only the 4th column (the CIDR pattern) is matched against the proxy-provided remote address override. Therefore, it is not possible to apply different rules to different client address when proxying TCP connections via a unix socket, because HBA rules for unix connections don't use the address column. Also when proxying client SSL connections via a non-SSL proxy connection, or proxying client non-SSL connections via a SSL proxy connection, care must be taken to configure address-based rule matching using the proper connection type. A reliable way to bypass this complexity is to only use the `host` connection type which applies equally to SSL and non-SSL connections. As of this implementation, the feature is enabled using the non-documented environment variable `COCKROACH_TRUST_CLIENT_PROVIDED_SQL_REMOTE_ADDR`. The use of an environment variable is a stop-gap so that this feature can be used in CockroachCloud SQL pods, which do not have access to [cluster settings](../v21.1/cluster-settings.html). The environment variable will be eventually removed and replaced by another mechanism. [#58381][#58381]
True
pgwire: make it possible for a SQL proxy to override the client addr - PR: https://github.com/cockroachdb/cockroach/pull/58381 When using a SQL proxy, in the default configuration CockroachDB only knows about the network address of the proxy. That *peer* address is then used for logging, [authentication](../v21.1/authentication.html) rules, etc. This is undesirable, as security logging and authentication rules need to operate on the actual (final) client address instead. CockroachDB can now be configured to solve this problem (conf mechanism detailed below). When so configured, a SQL proxy can inform the CockroachDB server of the real address of the client via a server status parameter called `crdb:remote_addr`. The value must be the IP address of the client, followed by a colon, followed by the port number, using the standard Go syntax (e.g., `11.22.33.44:5566` for IPv4, `[11:22::33]:4455` for IPv6). When provided, this value overrides the SQL proxy's address for logging and authentication purposes. In any case, the original peer address is also logged alongside the client address (overridden or not), via the new logging tag `peer`. Security considerations: - Enabling this feature allows the peer to spoof its address with regard to authentication and thus bypass authentication rules that would otherwise apply to its address, which can introduce a serious security vulnerability if the peer is not trusted. This is why this feature is not enabled by default, and must only be enabled when using a trusted SQL proxy. - This feature should only be used with SQL proxies which actively scrub a `crdb:remote_addr` parameter received by a remote client, and replaces it by its own. If the proxy mistakenly forwards the parameter as provided by the client, it opens the door to the aforementioned security vulnerability. - Care must be taken in HBA rules: TLS client cert validation, if requested by a rule, is still performed using the certificate presented by the proxy, not that presented by the client. This means that this new feature is not sufficient to forward TLS client cert authentication through a proxy. (If TLS client cert authentication is required, it must be performed by the proxy directly.) - Care must be taken in HBA rules: the `protocol` field (first column) continues to apply to the connection type between CockroachDB and the proxy, not between the proxy and the client. Only the 4th column (the CIDR pattern) is matched against the proxy-provided remote address override. Therefore, it is not possible to apply different rules to different client address when proxying TCP connections via a unix socket, because HBA rules for unix connections don't use the address column. Also when proxying client SSL connections via a non-SSL proxy connection, or proxying client non-SSL connections via a SSL proxy connection, care must be taken to configure address-based rule matching using the proper connection type. A reliable way to bypass this complexity is to only use the `host` connection type which applies equally to SSL and non-SSL connections. As of this implementation, the feature is enabled using the non-documented environment variable `COCKROACH_TRUST_CLIENT_PROVIDED_SQL_REMOTE_ADDR`. The use of an environment variable is a stop-gap so that this feature can be used in CockroachCloud SQL pods, which do not have access to [cluster settings](../v21.1/cluster-settings.html). The environment variable will be eventually removed and replaced by another mechanism. [#58381][#58381]
non_process
pgwire make it possible for a sql proxy to override the client addr pr when using a sql proxy in the default configuration cockroachdb only knows about the network address of the proxy that peer address is then used for logging authentication html rules etc this is undesirable as security logging and authentication rules need to operate on the actual final client address instead cockroachdb can now be configured to solve this problem conf mechanism detailed below when so configured a sql proxy can inform the cockroachdb server of the real address of the client via a server status parameter called crdb remote addr the value must be the ip address of the client followed by a colon followed by the port number using the standard go syntax e g for for when provided this value overrides the sql proxy s address for logging and authentication purposes in any case the original peer address is also logged alongside the client address overridden or not via the new logging tag peer security considerations enabling this feature allows the peer to spoof its address with regard to authentication and thus bypass authentication rules that would otherwise apply to its address which can introduce a serious security vulnerability if the peer is not trusted this is why this feature is not enabled by default and must only be enabled when using a trusted sql proxy this feature should only be used with sql proxies which actively scrub a crdb remote addr parameter received by a remote client and replaces it by its own if the proxy mistakenly forwards the parameter as provided by the client it opens the door to the aforementioned security vulnerability care must be taken in hba rules tls client cert validation if requested by a rule is still performed using the certificate presented by the proxy not that presented by the client this means that this new feature is not sufficient to forward tls client cert authentication through a proxy if tls client cert authentication is required it must be performed by the proxy directly care must be taken in hba rules the protocol field first column continues to apply to the connection type between cockroachdb and the proxy not between the proxy and the client only the column the cidr pattern is matched against the proxy provided remote address override therefore it is not possible to apply different rules to different client address when proxying tcp connections via a unix socket because hba rules for unix connections don t use the address column also when proxying client ssl connections via a non ssl proxy connection or proxying client non ssl connections via a ssl proxy connection care must be taken to configure address based rule matching using the proper connection type a reliable way to bypass this complexity is to only use the host connection type which applies equally to ssl and non ssl connections as of this implementation the feature is enabled using the non documented environment variable cockroach trust client provided sql remote addr the use of an environment variable is a stop gap so that this feature can be used in cockroachcloud sql pods which do not have access to cluster settings html the environment variable will be eventually removed and replaced by another mechanism
0
339,326
10,252,546,843
IssuesEvent
2019-08-21 09:14:58
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
app.pluralsight.com - video or audio doesn't play
browser-firefox engine-gecko priority-normal type-webrender-enabled
<!-- @browser: Firefox 70.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:70.0) Gecko/20100101 Firefox/70.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live **Browser / Version**: Firefox 70.0 **Operating System**: Mac OS X 10.14 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: opening a video, which opens a new tab to play it, does not start the video and instead gets stuck at loading **Steps to Reproduce**: 1. log into pluralsight 2. browse to video catalog 3. click to start course [![Screenshot Description](https://webcompat.com/uploads/2019/8/159f2d80-213c-40d4-ab4c-0b7ec9549367-thumb.jpeg)](https://webcompat.com/uploads/2019/8/159f2d80-213c-40d4-ab4c-0b7ec9549367.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190820093833</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [{u'log': [u'Request to access cookie or storage on https://cloud.typography.com/6966154/691568/css/fonts.css was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Loading failed for the <script> with source https://assets.adobedtm.com/2c5f2b30762ae62377916c09612cd9d60054fba9/satelliteLib-3f2001f88315610d295e482607f7d003c8627d08.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'2:1'}, {u'log': [u'onmozfullscreenchange is deprecated.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'12:18882'}, {u'log': [u'onmozfullscreenerror is deprecated.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'12:18882'}, {u'log': [u'Loading failed for the <script> with source https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Request to access cookie or storage on https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Loading failed for the <script> with source https://js-agent.newrelic.com/nr-974.min.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://webanalytics.pluralsight.com/web/webplayer_player_ui/identify. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'TypeError: NetworkError when attempting to fetch resource.'], u'level': u'error', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Synchronous XMLHttpRequest on the main thread is deprecated because of its detrimental effects to the end users experience. For more help http://xhr.spec.whatwg.org/'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'14:11294'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:04.682] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:04.683] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Request to access cookie or storage on https://cdn.segment.com/analytics.js/v1/xUu3kPNJl3xPuYRPGfTT5dmbdcGpXW39/analytics.min.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a.zxcvads.com/a/ads/-image-ad_ads.html. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/i. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/p. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/i. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://www.fullstory.com/s/fs.js. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Request to access cookie or storage on https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'<script> source URI is not allowed in this document: https://www.fullstory.com/s/fs.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Loading failed for the <script> with source https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'18:22095'}, {u'log': [u'NotAllowedError: The play method is not allowed by the user agent or the platform in the current context, possibly because the user denied permission.'], u'level': u'error', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:09.689] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:09.689] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:17.053] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:17.055] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:22.079] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:22.079] e:', u'Aborting failed request "/data". Max retries reached.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:22.079] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/m. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
app.pluralsight.com - video or audio doesn't play - <!-- @browser: Firefox 70.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:70.0) Gecko/20100101 Firefox/70.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live **Browser / Version**: Firefox 70.0 **Operating System**: Mac OS X 10.14 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: opening a video, which opens a new tab to play it, does not start the video and instead gets stuck at loading **Steps to Reproduce**: 1. log into pluralsight 2. browse to video catalog 3. click to start course [![Screenshot Description](https://webcompat.com/uploads/2019/8/159f2d80-213c-40d4-ab4c-0b7ec9549367-thumb.jpeg)](https://webcompat.com/uploads/2019/8/159f2d80-213c-40d4-ab4c-0b7ec9549367.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190820093833</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [{u'log': [u'Request to access cookie or storage on https://cloud.typography.com/6966154/691568/css/fonts.css was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Loading failed for the <script> with source https://assets.adobedtm.com/2c5f2b30762ae62377916c09612cd9d60054fba9/satelliteLib-3f2001f88315610d295e482607f7d003c8627d08.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'2:1'}, {u'log': [u'onmozfullscreenchange is deprecated.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'12:18882'}, {u'log': [u'onmozfullscreenerror is deprecated.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'12:18882'}, {u'log': [u'Loading failed for the <script> with source https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Request to access cookie or storage on https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Loading failed for the <script> with source https://js-agent.newrelic.com/nr-974.min.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://webanalytics.pluralsight.com/web/webplayer_player_ui/identify. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'TypeError: NetworkError when attempting to fetch resource.'], u'level': u'error', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Synchronous XMLHttpRequest on the main thread is deprecated because of its detrimental effects to the end users experience. For more help http://xhr.spec.whatwg.org/'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'14:11294'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:04.682] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:04.683] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Request to access cookie or storage on https://cdn.segment.com/analytics.js/v1/xUu3kPNJl3xPuYRPGfTT5dmbdcGpXW39/analytics.min.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a.zxcvads.com/a/ads/-image-ad_ads.html. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/i. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/p. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/i. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://www.fullstory.com/s/fs.js. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'Request to access cookie or storage on https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js was blocked because we are blocking all third-party storage access requests and content blocking is enabled.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'0:0'}, {u'log': [u'<script> source URI is not allowed in this document: https://www.fullstory.com/s/fs.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Loading failed for the <script> with source https://ssl.widgets.webengage.com/js/webengage-min-v-6.0.js.'], u'level': u'warn', u'uri': u'https://app.pluralsight.com/player?course=python-beyond-basics&author=austin-bingham&name=python-beyond-basics-m00&clip=0&mode=live', u'pos': u'1:1'}, {u'log': [u'Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted.'], u'level': u'warn', u'uri': u'https://s2.pluralsight.com/webplayer/player/client/libs/video.min.js', u'pos': u'18:22095'}, {u'log': [u'NotAllowedError: The play method is not allowed by the user agent or the platform in the current context, possibly because the user denied permission.'], u'level': u'error', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:09.689] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:09.689] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:17.053] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:17.055] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:22.079] e:', u'Fastdata request has failed.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'[Youbora][20:08:22.079] e:', u'Aborting failed request "/data". Max retries reached.'], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://a-fds.youborafds01.com/data?apiVersion=v7&outputformat=json&system=pluralsightprod&pluginVersion=6.5.11-adapterless&requestNumber=0.23102500261798575&timemark=1566335284625. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}, {u'log': [u'[Youbora][20:08:22.079] e:', {}], u'level': u'error', u'uri': u'https://s2.pluralsight.com/webplayer/js/main.3aa72d2825c5ddfe2b33.min.js', u'pos': u'106:5917'}, {u'log': [u'Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.segment.io/v1/m. (Reason: CORS request did not succeed).'], u'level': u'warn', u'uri': u'', u'pos': u'0:0'}] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
app pluralsight com video or audio doesn t play url browser version firefox operating system mac os x tested another browser yes problem type video or audio doesn t play description opening a video which opens a new tab to play it does not start the video and instead gets stuck at loading steps to reproduce log into pluralsight browse to video catalog click to start course browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled true gfx webrender all true channel nightly console messages u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u error u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log e u fastdata request has failed u level u error u uri u u pos u u log e u level u error u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u warn u uri u u pos u u log u level u error u uri u u pos u u log e u fastdata request has failed u level u error u uri u u pos u u log u level u warn u uri u u pos u u log e u level u error u uri u u pos u u log u level u warn u uri u u pos u u log e u fastdata request has failed u level u error u uri u u pos u u log e u level u error u uri u u pos u u log e u fastdata request has failed u level u error u uri u u pos u u log e u aborting failed request data max retries reached u level u error u uri u u pos u u log u level u warn u uri u u pos u u log e u level u error u uri u u pos u u log u level u warn u uri u u pos u from with ❤️
0
117,730
25,188,868,862
IssuesEvent
2022-11-11 21:18:17
creativecommons/search
https://api.github.com/repos/creativecommons/search
closed
[Feature] Improving the legibility of the site content to enhance readibility
🟩 priority: low 🏁 status: ready for work ✨ goal: improvement 💻 aspect: code
## Problem Website content should be legible to increase readability in the user experience. This is an important aspect that also contributes to users staying or leaving the website. I checked the website, and the first thing that caught my attention was the font size which was too small. This can prevent users with/without eye problems to read the content of the website. ## Description This is an image with the current font size which I think is not ideal. <img width="1053" alt="Screenshot 2022-10-14 at 9 51 06 AM" src="https://user-images.githubusercontent.com/70960505/195807693-4ca1f14b-032c-4a99-8d87-9acd514b68c3.png"> My suggestion is to increase the font size to 14px or 16px for paragraphs and for titles it can be 16px with font-weight of 500. ## Alternatives I also believe that using Arial or Roboto font family can improve the user experience. ## Implementation The implementation is quite simple: - the paragraph font size can be increased to 14px. - the title can be increased to 16px with a font weight of 500. <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in implementing this feature.
1.0
[Feature] Improving the legibility of the site content to enhance readibility - ## Problem Website content should be legible to increase readability in the user experience. This is an important aspect that also contributes to users staying or leaving the website. I checked the website, and the first thing that caught my attention was the font size which was too small. This can prevent users with/without eye problems to read the content of the website. ## Description This is an image with the current font size which I think is not ideal. <img width="1053" alt="Screenshot 2022-10-14 at 9 51 06 AM" src="https://user-images.githubusercontent.com/70960505/195807693-4ca1f14b-032c-4a99-8d87-9acd514b68c3.png"> My suggestion is to increase the font size to 14px or 16px for paragraphs and for titles it can be 16px with font-weight of 500. ## Alternatives I also believe that using Arial or Roboto font family can improve the user experience. ## Implementation The implementation is quite simple: - the paragraph font size can be increased to 14px. - the title can be increased to 16px with a font weight of 500. <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in implementing this feature.
non_process
improving the legibility of the site content to enhance readibility problem website content should be legible to increase readability in the user experience this is an important aspect that also contributes to users staying or leaving the website i checked the website and the first thing that caught my attention was the font size which was too small this can prevent users with without eye problems to read the content of the website description this is an image with the current font size which i think is not ideal img width alt screenshot at am src my suggestion is to increase the font size to or for paragraphs and for titles it can be with font weight of alternatives i also believe that using arial or roboto font family can improve the user experience implementation the implementation is quite simple the paragraph font size can be increased to the title can be increased to with a font weight of i would be interested in implementing this feature
0
225
2,652,285,897
IssuesEvent
2015-03-16 16:32:01
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
Traversal.reverse()
bug process
currently calls reverse() on each step. While that has the right semantics in some cases, i.e. ```java v.outE('knows').has('weight',5).inV ``` gets reversed to: ```java v.inE('knows').has('weight',5).outV ``` More complex traversals don't get reversed correctly: ```java v.outE('knows').has('weight',5).inV.outE('married').interval('since',0,10).inV ``` becomes ```java v.inE('knows').has('weight',5).outV.inE('married').interval('since',0,10).outV ``` whereas the "married" and "knows" should be flipped. Hence, more thought needs to be given to traversal reversal in the general case.
1.0
Traversal.reverse() - currently calls reverse() on each step. While that has the right semantics in some cases, i.e. ```java v.outE('knows').has('weight',5).inV ``` gets reversed to: ```java v.inE('knows').has('weight',5).outV ``` More complex traversals don't get reversed correctly: ```java v.outE('knows').has('weight',5).inV.outE('married').interval('since',0,10).inV ``` becomes ```java v.inE('knows').has('weight',5).outV.inE('married').interval('since',0,10).outV ``` whereas the "married" and "knows" should be flipped. Hence, more thought needs to be given to traversal reversal in the general case.
process
traversal reverse currently calls reverse on each step while that has the right semantics in some cases i e java v oute knows has weight inv gets reversed to java v ine knows has weight outv more complex traversals don t get reversed correctly java v oute knows has weight inv oute married interval since inv becomes java v ine knows has weight outv ine married interval since outv whereas the married and knows should be flipped hence more thought needs to be given to traversal reversal in the general case
1
11,785
14,617,069,271
IssuesEvent
2020-12-22 14:14:32
oj-tooth/lt_toolbox
https://api.github.com/repos/oj-tooth/lt_toolbox
closed
export_tracmass_to_nc
Post-Processing
* Write a function to export tracmass output data (.csv format) to a netCDF file.
1.0
export_tracmass_to_nc - * Write a function to export tracmass output data (.csv format) to a netCDF file.
process
export tracmass to nc write a function to export tracmass output data csv format to a netcdf file
1
4,187
7,134,601,348
IssuesEvent
2018-01-22 21:26:49
PHPSocialNetwork/phpfastcache
https://api.github.com/repos/PHPSocialNetwork/phpfastcache
closed
how to use redis when my hosting provider use a sock for the connexion
6.0 6.1 7.0 >_< Working & Scheduled [-_-] In Process ~_~ Issue confirmed
### Configuration: PhpFastCache version: 6.0.8 PHP version: 7.1 Operating system: Apache ??? #### Issue description: Hi, i know that my hosting provider tell us that for using redis, we have to connect with a socket like /var/run/redis/redis.sock i tried to do that : $InstanceCache = CacheManager::getInstance('redis', [ 'host' => '/var/run/redis/redis.sock' ]); but i had a fatal error PHP message: PHP Warning: Redis::connect(): php_network_getaddresses: getaddrinfo failed: Name or service not known in /var/www/calendriergratuit.fr/www/vendor/phpfastcache/phpfastcache/src/phpFastCache/Drivers/Redis/Driver.php on line 137 How to configure redis driver in my case?
1.0
how to use redis when my hosting provider use a sock for the connexion - ### Configuration: PhpFastCache version: 6.0.8 PHP version: 7.1 Operating system: Apache ??? #### Issue description: Hi, i know that my hosting provider tell us that for using redis, we have to connect with a socket like /var/run/redis/redis.sock i tried to do that : $InstanceCache = CacheManager::getInstance('redis', [ 'host' => '/var/run/redis/redis.sock' ]); but i had a fatal error PHP message: PHP Warning: Redis::connect(): php_network_getaddresses: getaddrinfo failed: Name or service not known in /var/www/calendriergratuit.fr/www/vendor/phpfastcache/phpfastcache/src/phpFastCache/Drivers/Redis/Driver.php on line 137 How to configure redis driver in my case?
process
how to use redis when my hosting provider use a sock for the connexion configuration phpfastcache version php version operating system apache issue description hi i know that my hosting provider tell us that for using redis we have to connect with a socket like var run redis redis sock i tried to do that instancecache cachemanager getinstance redis host var run redis redis sock but i had a fatal error php message php warning redis connect php network getaddresses getaddrinfo failed name or service not known in var www calendriergratuit fr www vendor phpfastcache phpfastcache src phpfastcache drivers redis driver php on line how to configure redis driver in my case
1
63,462
26,406,256,491
IssuesEvent
2023-01-13 08:17:44
LiskHQ/lisk-service
https://api.github.com/repos/LiskHQ/lisk-service
opened
Remove rank calculation logic
type: improvement service/blockchain-connector
### Description The indexer currently computes the validator rank explicitly. With Lisk SDK `v6.0.0-alpha.11`, the `pos_getValidatorsByStake` endpoint can now return all the validators ordered by stake when invoked with the limit parameter `-1`. ### Acceptance Criteria - `getAllPoSValidators` in the connector should now fetch all the validators using the `pos_getValidatorsByStake` endpoint - The rank computation logic is removed - Validator `rank` is assigned based on their index in the response from `pos_getValidatorsByStake`
1.0
Remove rank calculation logic - ### Description The indexer currently computes the validator rank explicitly. With Lisk SDK `v6.0.0-alpha.11`, the `pos_getValidatorsByStake` endpoint can now return all the validators ordered by stake when invoked with the limit parameter `-1`. ### Acceptance Criteria - `getAllPoSValidators` in the connector should now fetch all the validators using the `pos_getValidatorsByStake` endpoint - The rank computation logic is removed - Validator `rank` is assigned based on their index in the response from `pos_getValidatorsByStake`
non_process
remove rank calculation logic description the indexer currently computes the validator rank explicitly with lisk sdk alpha the pos getvalidatorsbystake endpoint can now return all the validators ordered by stake when invoked with the limit parameter acceptance criteria getallposvalidators in the connector should now fetch all the validators using the pos getvalidatorsbystake endpoint the rank computation logic is removed validator rank is assigned based on their index in the response from pos getvalidatorsbystake
0
683,463
23,383,321,579
IssuesEvent
2022-08-11 11:36:37
CA-G12/games-proj-Tariq-shams
https://api.github.com/repos/CA-G12/games-proj-Tariq-shams
closed
[HTML, CSS, JS] Create the file structure
in-progress priority-0
## We will add the following: - Public folder that has index.html and a js folder including the dom & the fetch files, in addition to the css file. - Src folder including router.js, server.js, and test.js. - NPM packages: nodemon, jest, and ESLint. - Data.json file.
1.0
[HTML, CSS, JS] Create the file structure - ## We will add the following: - Public folder that has index.html and a js folder including the dom & the fetch files, in addition to the css file. - Src folder including router.js, server.js, and test.js. - NPM packages: nodemon, jest, and ESLint. - Data.json file.
non_process
create the file structure we will add the following public folder that has index html and a js folder including the dom the fetch files in addition to the css file src folder including router js server js and test js npm packages nodemon jest and eslint data json file
0
136,954
5,290,803,892
IssuesEvent
2017-02-08 20:50:51
urbit/urbit
https://api.github.com/repos/urbit/urbit
closed
Line noise on startup, never ends.
%ames difficulty high priority low
I'm getting a lot of noise when I start urbit via 'urbit hidnyl-matped' on Mac OS. My version of urbit is up to date from git-head (installed via homebrew, though the issue persists when built manually). It never actually stops spewing the noise, but I can use ^V to get to the process manager. Please excuse the volume of text: urbit: home is hidnyl-matped loom: mapped 2048MB protected loom live: loaded: MB/138.543.104 boot: installed 229 jets arvo: time: ~2015.10.8..11.20.36..ae38 rest: checkpoint to event 29.681 rest: old 0v1h.1obov, new 0v1r.43kv3 loaded passcode from hidnyl-matped/.urb/code.~simdem-fosnep ---------------- playback complete---------------- ames: on localhost, UDP 52771. http: live (insecure) on 8080 http: live ("secure") on 8443 term: live (but unsupported) on 10023 [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.226 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.227 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.228 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.230 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 11].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[419 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[420 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[423 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[424 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[425 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[426 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[427 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[435 13].[465 53]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[436 15].[456 17]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[438 21].[438 71]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[562 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 11].[563 18]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[570 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[571 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[586 9].[586 21]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[574 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 13].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[576 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[577 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[580 11].[583 13]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 22].[582 49]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 24].[582 49]> \/ \/ need [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.226 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.227 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.228 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.230 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 11].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[419 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[420 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[423 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[424 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[425 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[426 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[427 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[435 13].[465 53]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[436 15].[456 17]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[438 21].[438 71]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[562 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 11].[563 18]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[570 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[571 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[586 9].[586 21]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[574 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 13].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[576 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[577 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[580 11].[583 13]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 22].[582 49]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 24].[582 49]> \/ \/ need Then after a while, I get the following, which I assume is basically a time-out and reattempt. kick: lost %hear on //ames ames: czar zod.urbit.org: ip .192.241.195.84 [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/
1.0
Line noise on startup, never ends. - I'm getting a lot of noise when I start urbit via 'urbit hidnyl-matped' on Mac OS. My version of urbit is up to date from git-head (installed via homebrew, though the issue persists when built manually). It never actually stops spewing the noise, but I can use ^V to get to the process manager. Please excuse the volume of text: urbit: home is hidnyl-matped loom: mapped 2048MB protected loom live: loaded: MB/138.543.104 boot: installed 229 jets arvo: time: ~2015.10.8..11.20.36..ae38 rest: checkpoint to event 29.681 rest: old 0v1h.1obov, new 0v1r.43kv3 loaded passcode from hidnyl-matped/.urb/code.~simdem-fosnep ---------------- playback complete---------------- ames: on localhost, UDP 52771. http: live (insecure) on 8080 http: live ("secure") on 8443 term: live (but unsupported) on 10023 [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.226 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.227 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.228 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.230 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 11].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[419 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[420 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[423 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[424 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[425 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[426 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[427 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[435 13].[465 53]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[436 15].[456 17]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[438 21].[438 71]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[562 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 11].[563 18]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[570 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[571 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[586 9].[586 21]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[574 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 13].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[576 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[577 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[580 11].[583 13]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 22].[582 49]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 24].[582 49]> \/ \/ need [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.226 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.227 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.228 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.230 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.231 11].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[419 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[420 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[423 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[424 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[425 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[426 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[427 5].[466 7]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[435 13].[465 53]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[436 15].[456 17]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[438 21].[438 71]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[562 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 7].[567 9]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[563 11].[563 18]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[570 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[571 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[586 9].[586 21]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[585 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[587 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[588 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[572 11].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[573 7].[588 60]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[574 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[575 13].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[576 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[577 9].[584 11]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[580 11].[583 13]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 22].[582 49]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[582 24].[582 49]> \/ \/ need Then after a while, I get the following, which I assume is basically a time-out and reattempt. kick: lost %hear on //ames ames: czar zod.urbit.org: ip .192.241.195.84 [%ap-fill-full [~hidnyl-matped %talk] ~doznec 2] [%swim-take-vane %g %made ~] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.223 3].[1.231 36]> \/ \/ \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.224 3].[1.231 36]> \/ \/ [ %gall-take / ~hidnyl-matped use talk ~doznec out talk friend show porch ~doznec urbit-meta ] \//~ramrym-waclyt-botsur-wanhus--darfel-bilnel-rolduc-tomden/home/~2015.10.6\/ ..18.31.22..427e/arvo/gall/:<[1.225 3].[1.231 36]> \/
non_process
line noise on startup never ends i m getting a lot of noise when i start urbit via urbit hidnyl matped on mac os my version of urbit is up to date from git head installed via homebrew though the issue persists when built manually it never actually stops spewing the noise but i can use v to get to the process manager please excuse the volume of text urbit home is hidnyl matped loom mapped protected loom live loaded mb boot installed jets arvo time rest checkpoint to event rest old new loaded passcode from hidnyl matped urb code simdem fosnep playback complete ames on localhost udp http live insecure on http live secure on term live but unsupported on doznec ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall gall take hidnyl matped use talk doznec out talk friend show porch doznec urbit meta ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall need doznec ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall gall take hidnyl matped use talk doznec out talk friend show porch doznec urbit meta ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall need then after a while i get the following which i assume is basically a time out and reattempt kick lost hear on ames ames czar zod urbit org ip doznec ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall gall take hidnyl matped use talk doznec out talk friend show porch doznec urbit meta ramrym waclyt botsur wanhus darfel bilnel rolduc tomden home arvo gall
0
6,436
9,538,701,633
IssuesEvent
2019-04-30 15:15:35
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
[TextFields] Explore removing notion of error state from MDCInputTextField
[TextFields] type:Process
This was filed as an internal issue. If you are a Googler, please visit [b/131170010](http://b/131170010) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/131170010](http://b/131170010)
1.0
[TextFields] Explore removing notion of error state from MDCInputTextField - This was filed as an internal issue. If you are a Googler, please visit [b/131170010](http://b/131170010) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/131170010](http://b/131170010)
process
explore removing notion of error state from mdcinputtextfield this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug
1
288,449
31,861,376,273
IssuesEvent
2023-09-15 11:10:33
nidhi7598/linux-v4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-v4.19.72_CVE-2022-3564
opened
WS-2021-0232 (High) detected in linuxlinux-4.19.294
Mend: dependency security vulnerability
## WS-2021-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/caif/caif_usb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/caif/caif_usb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux/Kernel in versions v3.4-rc1 to v5.12.9 is vulnerable to memory leak in cfusbl_device_notify <p>Publish Date: 2021-06-04 <p>URL: <a href=https://github.com/gregkh/linux/commit/4d94f530cd24c85aede6e72b8923f371b45d6886>WS-2021-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000750">https://osv.dev/vulnerability/UVI-2021-1000750</a></p> <p>Release Date: 2021-06-04</p> <p>Fix Resolution: v5.4.125</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0232 (High) detected in linuxlinux-4.19.294 - ## WS-2021-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/caif/caif_usb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/caif/caif_usb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux/Kernel in versions v3.4-rc1 to v5.12.9 is vulnerable to memory leak in cfusbl_device_notify <p>Publish Date: 2021-06-04 <p>URL: <a href=https://github.com/gregkh/linux/commit/4d94f530cd24c85aede6e72b8923f371b45d6886>WS-2021-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000750">https://osv.dev/vulnerability/UVI-2021-1000750</a></p> <p>Release Date: 2021-06-04</p> <p>Fix Resolution: v5.4.125</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in linuxlinux ws high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files net caif caif usb c net caif caif usb c vulnerability details linux kernel in versions to is vulnerable to memory leak in cfusbl device notify publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
3,732
6,733,142,433
IssuesEvent
2017-10-18 13:58:38
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
Contract preparation single item view - Data Loader
process workflow
Initialize and load data collections and page items.
1.0
Contract preparation single item view - Data Loader - Initialize and load data collections and page items.
process
contract preparation single item view data loader initialize and load data collections and page items
1
21,941
30,446,799,413
IssuesEvent
2023-07-15 19:28:42
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpgwzk_zri/pyutils" } }```
1.0
pyutils 0.0.1 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpgwzk_zri/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpgwzk zri pyutils
1
416,982
12,153,102,279
IssuesEvent
2020-04-25 00:40:00
Azure/ARO-RP
https://api.github.com/repos/Azure/ARO-RP
closed
Return deployment errors to end user?
priority-high size-small
Investigate to see if we can pass through deployment errors to end user and not whitelist them on a case by case basis. Example - so they can see errors like ZonalAllocationFailure, RequestDisallowedByPolicy, etc. These are in principle visible to end users via the CRG anyway.
1.0
Return deployment errors to end user? - Investigate to see if we can pass through deployment errors to end user and not whitelist them on a case by case basis. Example - so they can see errors like ZonalAllocationFailure, RequestDisallowedByPolicy, etc. These are in principle visible to end users via the CRG anyway.
non_process
return deployment errors to end user investigate to see if we can pass through deployment errors to end user and not whitelist them on a case by case basis example so they can see errors like zonalallocationfailure requestdisallowedbypolicy etc these are in principle visible to end users via the crg anyway
0
283,508
30,913,326,333
IssuesEvent
2023-08-05 01:40:25
hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
reopened
CVE-2021-28964 (Medium) detected in linuxlinux-4.19.279
Mend: dependency security vulnerability
## CVE-2021-28964 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/ctree.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A race condition was discovered in get_old_root in fs/btrfs/ctree.c in the Linux kernel through 5.11.8. It allows attackers to cause a denial of service (BUG) because of a lack of locking on an extent buffer before a cloning operation, aka CID-dbcc7d57bffc. <p>Publish Date: 2021-03-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28964>CVE-2021-28964</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-28964">https://www.linuxkernelcves.com/cves/CVE-2021-28964</a></p> <p>Release Date: 2021-03-22</p> <p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-28964 (Medium) detected in linuxlinux-4.19.279 - ## CVE-2021-28964 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/ctree.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A race condition was discovered in get_old_root in fs/btrfs/ctree.c in the Linux kernel through 5.11.8. It allows attackers to cause a denial of service (BUG) because of a lack of locking on an extent buffer before a cloning operation, aka CID-dbcc7d57bffc. <p>Publish Date: 2021-03-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-28964>CVE-2021-28964</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-28964">https://www.linuxkernelcves.com/cves/CVE-2021-28964</a></p> <p>Release Date: 2021-03-22</p> <p>Fix Resolution: v4.4.263, v4.9.263, v4.14.227, v4.19.183, v5.4.108, v5.10.26, v5.11.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files fs btrfs ctree c vulnerability details a race condition was discovered in get old root in fs btrfs ctree c in the linux kernel through it allows attackers to cause a denial of service bug because of a lack of locking on an extent buffer before a cloning operation aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
295,848
25,508,818,933
IssuesEvent
2022-11-28 11:31:13
ArielHeleneto/RISCV-testcase
https://api.github.com/repos/ArielHeleneto/RISCV-testcase
opened
[Add Testcase] DDE
Testcase / 测试用例
### 请选择: - [X] 我已经检查了 [Issue](https://github.com/brsf11/mugen-riscv/issues),进行了搜索,但没有发现类似问题。 - [X] 我正在着手整理该测试用例库。 - [X] 我希望接手该测试用例库。 ### 软件名称 DDE ### 软件版本 DDE ### 操作系统版本 openEuler 22.03 LTS riscv64 ### 软件描述 DDE (Deepin Desktop Enviroment) 是指Deepin所搭载的中国自主桌面环境。 ### 已有测试用例来源 目前已有的测试用例: ### 其他有助于解决该问题的内容 无
1.0
[Add Testcase] DDE - ### 请选择: - [X] 我已经检查了 [Issue](https://github.com/brsf11/mugen-riscv/issues),进行了搜索,但没有发现类似问题。 - [X] 我正在着手整理该测试用例库。 - [X] 我希望接手该测试用例库。 ### 软件名称 DDE ### 软件版本 DDE ### 操作系统版本 openEuler 22.03 LTS riscv64 ### 软件描述 DDE (Deepin Desktop Enviroment) 是指Deepin所搭载的中国自主桌面环境。 ### 已有测试用例来源 目前已有的测试用例: ### 其他有助于解决该问题的内容 无
non_process
dde 请选择: 我已经检查了 我正在着手整理该测试用例库。 我希望接手该测试用例库。 软件名称 dde 软件版本 dde 操作系统版本 openeuler lts 软件描述 dde deepin desktop enviroment 是指deepin所搭载的中国自主桌面环境。 已有测试用例来源 目前已有的测试用例: 其他有助于解决该问题的内容 无
0
2,542
5,300,554,819
IssuesEvent
2017-02-10 05:38:36
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Retry failed uploads with Atlas post-processor
enhancement post-1.0 post-processor/atlas
Sometimes an upload to Atlas fails because of a network hiccup or other transient error. Since the build completed successfully we should try again to upload it. Packer already has retry wrappers all over the place so I'm surprised that this does not have one already, but we should be able to copy some code over.
1.0
Retry failed uploads with Atlas post-processor - Sometimes an upload to Atlas fails because of a network hiccup or other transient error. Since the build completed successfully we should try again to upload it. Packer already has retry wrappers all over the place so I'm surprised that this does not have one already, but we should be able to copy some code over.
process
retry failed uploads with atlas post processor sometimes an upload to atlas fails because of a network hiccup or other transient error since the build completed successfully we should try again to upload it packer already has retry wrappers all over the place so i m surprised that this does not have one already but we should be able to copy some code over
1
3,048
6,042,164,752
IssuesEvent
2017-06-11 10:13:57
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Assign className to clicked row on expand
help wanted inprocess
Is there a way to add a custom background color or className to a row when a user clicks to expand it, in the same way that you can with the select row option? I just want the row to expand, and not be selectable.
1.0
Assign className to clicked row on expand - Is there a way to add a custom background color or className to a row when a user clicks to expand it, in the same way that you can with the select row option? I just want the row to expand, and not be selectable.
process
assign classname to clicked row on expand is there a way to add a custom background color or classname to a row when a user clicks to expand it in the same way that you can with the select row option i just want the row to expand and not be selectable
1
11,278
14,077,950,381
IssuesEvent
2020-11-04 12:51:13
tommy-josepovic/smarthome-simulator-team-2
https://api.github.com/repos/tommy-josepovic/smarthome-simulator-team-2
opened
D2 - Use Cases
software process
Use cases for the following features: Simulation parameters - System shall allow simulator’s user to grant/deny permissions for executing commands on the registered smart home modules - System shall save in a file the profiles, with their corresponding permissions, to avoid reentering the information every time you open the simulator Context of the simulation - System shall allow the simulator’s user to change the time speed Smart home core functionality (SHC) module - System shall allow users /smart home modules to open/close doors - System shall allow users /smart home modules to open/close windows - System shall allow users /smart home modules to turn on/off lights - System shall allow users to set an Auto mode, that turn on/off lights automatically when someone enters to a room Smart home security (SHP) Module - System shall allow users to set an away mode - System shall send notification to users if any motion detectors are triggered while the user is away - System shall allow users to set how much time should pass before alerting the authorities, when motion detectors are triggered during set away mode Smart home dashboard - System shall update the house view to graphically represent the current state of the simulation, i.e., when the simulation is activated (on-toggle button)
1.0
D2 - Use Cases - Use cases for the following features: Simulation parameters - System shall allow simulator’s user to grant/deny permissions for executing commands on the registered smart home modules - System shall save in a file the profiles, with their corresponding permissions, to avoid reentering the information every time you open the simulator Context of the simulation - System shall allow the simulator’s user to change the time speed Smart home core functionality (SHC) module - System shall allow users /smart home modules to open/close doors - System shall allow users /smart home modules to open/close windows - System shall allow users /smart home modules to turn on/off lights - System shall allow users to set an Auto mode, that turn on/off lights automatically when someone enters to a room Smart home security (SHP) Module - System shall allow users to set an away mode - System shall send notification to users if any motion detectors are triggered while the user is away - System shall allow users to set how much time should pass before alerting the authorities, when motion detectors are triggered during set away mode Smart home dashboard - System shall update the house view to graphically represent the current state of the simulation, i.e., when the simulation is activated (on-toggle button)
process
use cases use cases for the following features simulation parameters system shall allow simulator’s user to grant deny permissions for executing commands on the registered smart home modules system shall save in a file the profiles with their corresponding permissions to avoid reentering the information every time you open the simulator context of the simulation system shall allow the simulator’s user to change the time speed smart home core functionality shc module system shall allow users smart home modules to open close doors system shall allow users smart home modules to open close windows system shall allow users smart home modules to turn on off lights system shall allow users to set an auto mode that turn on off lights automatically when someone enters to a room smart home security shp module system shall allow users to set an away mode system shall send notification to users if any motion detectors are triggered while the user is away system shall allow users to set how much time should pass before alerting the authorities when motion detectors are triggered during set away mode smart home dashboard system shall update the house view to graphically represent the current state of the simulation i e when the simulation is activated on toggle button
1
16,418
21,212,966,335
IssuesEvent
2022-04-11 02:39:03
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
closed
Step-by-step program execution
enhancement processor
For debugging purposes, it would be great to have a way to execute a program one cycle at a time, and to be able to read the state of the VM at every cycle. One way to do this is described below: First, we could create a separate function in the `processor` crate (where the `execute()` function is defined) to handle step-by-step execution. This function could be called `execute_iter()` and it would return an iterator over process states (we'll also need to define a struct which represents a process state). Internally, `execute_iter()` could do the same thing as `execute()` but instead of converting a `Process` into an `ExecutionTrace`, it would iterate over the data contained in the process and return it for a given cycle. Basically, we would execute the program to the end (or until there is an error), and then start reading generated states one-by-one as a part of iteration process. This would be relatively "wasteful", but I think that for debug purposes its fine. As mentioned above, the `Process` struct already contains most of the historical data that we'd need (e.g., it keeps track of how memory changed with every cycle). There are two things for which it keeps only the latest state (rather than all historical states): * Stack overflow table. * Advice provider. I think we can tackle advice provider part at a later time, but it would be good to come up with a way to keep historical states of the stack overflow table. This will probably be relatively heavy - so, this functionality should be optional. Meaning, we'd use it only when executing programs from `execute_iter()` context, but not when using the regular `execute()` context.
1.0
Step-by-step program execution - For debugging purposes, it would be great to have a way to execute a program one cycle at a time, and to be able to read the state of the VM at every cycle. One way to do this is described below: First, we could create a separate function in the `processor` crate (where the `execute()` function is defined) to handle step-by-step execution. This function could be called `execute_iter()` and it would return an iterator over process states (we'll also need to define a struct which represents a process state). Internally, `execute_iter()` could do the same thing as `execute()` but instead of converting a `Process` into an `ExecutionTrace`, it would iterate over the data contained in the process and return it for a given cycle. Basically, we would execute the program to the end (or until there is an error), and then start reading generated states one-by-one as a part of iteration process. This would be relatively "wasteful", but I think that for debug purposes its fine. As mentioned above, the `Process` struct already contains most of the historical data that we'd need (e.g., it keeps track of how memory changed with every cycle). There are two things for which it keeps only the latest state (rather than all historical states): * Stack overflow table. * Advice provider. I think we can tackle advice provider part at a later time, but it would be good to come up with a way to keep historical states of the stack overflow table. This will probably be relatively heavy - so, this functionality should be optional. Meaning, we'd use it only when executing programs from `execute_iter()` context, but not when using the regular `execute()` context.
process
step by step program execution for debugging purposes it would be great to have a way to execute a program one cycle at a time and to be able to read the state of the vm at every cycle one way to do this is described below first we could create a separate function in the processor crate where the execute function is defined to handle step by step execution this function could be called execute iter and it would return an iterator over process states we ll also need to define a struct which represents a process state internally execute iter could do the same thing as execute but instead of converting a process into an executiontrace it would iterate over the data contained in the process and return it for a given cycle basically we would execute the program to the end or until there is an error and then start reading generated states one by one as a part of iteration process this would be relatively wasteful but i think that for debug purposes its fine as mentioned above the process struct already contains most of the historical data that we d need e g it keeps track of how memory changed with every cycle there are two things for which it keeps only the latest state rather than all historical states stack overflow table advice provider i think we can tackle advice provider part at a later time but it would be good to come up with a way to keep historical states of the stack overflow table this will probably be relatively heavy so this functionality should be optional meaning we d use it only when executing programs from execute iter context but not when using the regular execute context
1
749,191
26,153,104,180
IssuesEvent
2022-12-30 16:55:02
chaotic-aur/packages
https://api.github.com/repos/chaotic-aur/packages
closed
[Request] libreoffice-dev-bin
request:new-pkg priority:low
### Link to the package base(s) in the AUR https://aur.archlinux.org/packages/libreoffice-dev-bin ### Utility this package has for you _No response_ ### Do you consider the package(s) to be useful for every Chaotic-AUR user? YES! ### Do you consider the package to be useful for feature testing/preview? - [X] Yes ### Have you tested if the package builds in a clean chroot? - [X] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information replace this _LOver=7.5.0.1 pkgver=7.5.0.1 _pkgnamefmt=LibreOffice
1.0
[Request] libreoffice-dev-bin - ### Link to the package base(s) in the AUR https://aur.archlinux.org/packages/libreoffice-dev-bin ### Utility this package has for you _No response_ ### Do you consider the package(s) to be useful for every Chaotic-AUR user? YES! ### Do you consider the package to be useful for feature testing/preview? - [X] Yes ### Have you tested if the package builds in a clean chroot? - [X] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information replace this _LOver=7.5.0.1 pkgver=7.5.0.1 _pkgnamefmt=LibreOffice
non_process
libreoffice dev bin link to the package base s in the aur utility this package has for you no response do you consider the package s to be useful for every chaotic aur user yes do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information replace this lover pkgver pkgnamefmt libreoffice
0
256,922
22,112,089,186
IssuesEvent
2022-06-01 22:18:10
astropy/astroquery
https://api.github.com/repos/astropy/astroquery
opened
ENH TST: consider using pytest-timeout
enhancement testing
While we have the timeout arg/config for most modules, something things still stuck, therefore we should consider dealing with it on the test infra level e.g. with pytest-timeout. I believe sunpy uses it, maybe @Cadair can share the tricks if any of them are required.
1.0
ENH TST: consider using pytest-timeout - While we have the timeout arg/config for most modules, something things still stuck, therefore we should consider dealing with it on the test infra level e.g. with pytest-timeout. I believe sunpy uses it, maybe @Cadair can share the tricks if any of them are required.
non_process
enh tst consider using pytest timeout while we have the timeout arg config for most modules something things still stuck therefore we should consider dealing with it on the test infra level e g with pytest timeout i believe sunpy uses it maybe cadair can share the tricks if any of them are required
0
15,184
18,955,117,349
IssuesEvent
2021-11-18 19:17:40
Arch666Angel/mods
https://api.github.com/repos/Arch666Angel/mods
closed
Bio Token issues
Angels Bio Processing
## Issues - When playing with a high tech cost modifier, unlocking token / garden duplication requires a prohibitively high number of tokens - Many techs that feel like they should require tokens actually don't - No farming recipes require T2 farms, only the garden duplication recipes ## Suggested fixes - Add token to techs: - Biters x3 - Puffers x4 - Pressing x2 - Fish Pressing x2 - Butchery 2 - Alien Processing 2 - Alien Processing 3 - Crystals x6 - Fermentation - Bio Plastic x2 - Move garden duplication recipes to Alien Farming - Make fish accessible earlier in the tech tree - Remove Water treatment 3 as a prerequisite for Fish refugium 1 (presumably for Saline water but this is already accessible from Hydro plant or Washing plant) - Make Artificial fish water craftable in a regular Chemical plant (so no prerequisite of Advanced chemistry 1 is required) - Swap environment and T2 farming techs - Make T1 farming techs ignore tech cost modifier - Make garden duplication recipes craftable in T1 farm - Make T2 farming recipes require T2 farms
1.0
Bio Token issues - ## Issues - When playing with a high tech cost modifier, unlocking token / garden duplication requires a prohibitively high number of tokens - Many techs that feel like they should require tokens actually don't - No farming recipes require T2 farms, only the garden duplication recipes ## Suggested fixes - Add token to techs: - Biters x3 - Puffers x4 - Pressing x2 - Fish Pressing x2 - Butchery 2 - Alien Processing 2 - Alien Processing 3 - Crystals x6 - Fermentation - Bio Plastic x2 - Move garden duplication recipes to Alien Farming - Make fish accessible earlier in the tech tree - Remove Water treatment 3 as a prerequisite for Fish refugium 1 (presumably for Saline water but this is already accessible from Hydro plant or Washing plant) - Make Artificial fish water craftable in a regular Chemical plant (so no prerequisite of Advanced chemistry 1 is required) - Swap environment and T2 farming techs - Make T1 farming techs ignore tech cost modifier - Make garden duplication recipes craftable in T1 farm - Make T2 farming recipes require T2 farms
process
bio token issues issues when playing with a high tech cost modifier unlocking token garden duplication requires a prohibitively high number of tokens many techs that feel like they should require tokens actually don t no farming recipes require farms only the garden duplication recipes suggested fixes add token to techs biters puffers pressing fish pressing butchery alien processing alien processing crystals fermentation bio plastic move garden duplication recipes to alien farming make fish accessible earlier in the tech tree remove water treatment as a prerequisite for fish refugium presumably for saline water but this is already accessible from hydro plant or washing plant make artificial fish water craftable in a regular chemical plant so no prerequisite of advanced chemistry is required swap environment and farming techs make farming techs ignore tech cost modifier make garden duplication recipes craftable in farm make farming recipes require farms
1
297,713
25,758,208,549
IssuesEvent
2022-12-08 18:06:20
CSC207-2022F-UofT/course-project-team-tree-dog
https://api.github.com/repos/CSC207-2022F-UofT/course-project-team-tree-dog
closed
[Unit Test] Thread Lock Architecture
Test
- Write a test suite which tests all use cases as they interact with each other! Start multiple use case threads, think through the possible expected outcomes that can occur depending on the order in which each thread catches locks first (wins the lock race), add randomized sleeps and thread start order to ensure that every scenario can occur, and assert that only one of the expected scenarios occur. - Set `fairness = true` for all Reentrant locks created across the project. this ensures that locks will give priority to threads that have been waiting the longest, ensuring that every thread gets their chance to take the lock and use the resource. - See Bracket diagram in Conventions doc. There is a potential rare deadlock where JPL acquires its own "lock" and SP acquires "pool" and "game". Now JPL is waiting on "pool" and SP is waiting on "lock". This occurrence is very rare since each JPL thread has its own lock, so the "lock" would have to match for the same JPL for the deadlock to occur, and of course, the race conditions have to have a specific outcome. Still, this is a possible issue and should be fixed.
1.0
[Unit Test] Thread Lock Architecture - - Write a test suite which tests all use cases as they interact with each other! Start multiple use case threads, think through the possible expected outcomes that can occur depending on the order in which each thread catches locks first (wins the lock race), add randomized sleeps and thread start order to ensure that every scenario can occur, and assert that only one of the expected scenarios occur. - Set `fairness = true` for all Reentrant locks created across the project. this ensures that locks will give priority to threads that have been waiting the longest, ensuring that every thread gets their chance to take the lock and use the resource. - See Bracket diagram in Conventions doc. There is a potential rare deadlock where JPL acquires its own "lock" and SP acquires "pool" and "game". Now JPL is waiting on "pool" and SP is waiting on "lock". This occurrence is very rare since each JPL thread has its own lock, so the "lock" would have to match for the same JPL for the deadlock to occur, and of course, the race conditions have to have a specific outcome. Still, this is a possible issue and should be fixed.
non_process
thread lock architecture write a test suite which tests all use cases as they interact with each other start multiple use case threads think through the possible expected outcomes that can occur depending on the order in which each thread catches locks first wins the lock race add randomized sleeps and thread start order to ensure that every scenario can occur and assert that only one of the expected scenarios occur set fairness true for all reentrant locks created across the project this ensures that locks will give priority to threads that have been waiting the longest ensuring that every thread gets their chance to take the lock and use the resource see bracket diagram in conventions doc there is a potential rare deadlock where jpl acquires its own lock and sp acquires pool and game now jpl is waiting on pool and sp is waiting on lock this occurrence is very rare since each jpl thread has its own lock so the lock would have to match for the same jpl for the deadlock to occur and of course the race conditions have to have a specific outcome still this is a possible issue and should be fixed
0
17,520
23,329,290,008
IssuesEvent
2022-08-09 02:18:59
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[Feature][FLINK-28082] Support end to end encryption on connector
compute/data-processing type/feature
Pulsar support [end-to-end encryption](https://pulsar.apache.org/docs/en/security-encryption/#get-started). We need to add this ability to the Pulsar connector by supporting `CryptoKeyReader`. - [x] Implement `CryptoKeyReader` on source. - [x] Implement `CryptoKeyReader` on sink.
1.0
[Feature][FLINK-28082] Support end to end encryption on connector - Pulsar support [end-to-end encryption](https://pulsar.apache.org/docs/en/security-encryption/#get-started). We need to add this ability to the Pulsar connector by supporting `CryptoKeyReader`. - [x] Implement `CryptoKeyReader` on source. - [x] Implement `CryptoKeyReader` on sink.
process
support end to end encryption on connector pulsar support we need to add this ability to the pulsar connector by supporting cryptokeyreader implement cryptokeyreader on source implement cryptokeyreader on sink
1
657
3,126,742,796
IssuesEvent
2015-09-08 11:08:19
vasi/pixz
https://api.github.com/repos/vasi/pixz
closed
1.0.3 tarball not on SourceForge
release process
It looks like pixz 1.0.3 was released, but its tarball was not uploaded to [SourceForge](http://sourceforge.net/projects/pixz/files/) where the pixz github project says the file should be.
1.0
1.0.3 tarball not on SourceForge - It looks like pixz 1.0.3 was released, but its tarball was not uploaded to [SourceForge](http://sourceforge.net/projects/pixz/files/) where the pixz github project says the file should be.
process
tarball not on sourceforge it looks like pixz was released but its tarball was not uploaded to where the pixz github project says the file should be
1
5,041
7,857,345,270
IssuesEvent
2018-06-21 10:27:46
joostverdoorn/webrtc
https://api.github.com/repos/joostverdoorn/webrtc
closed
Als afstudeerder, wil ik een blogpost over de interessante ontdekkingen van het project doen
process
<!--- @huboard:{"order":105.0} -->
1.0
Als afstudeerder, wil ik een blogpost over de interessante ontdekkingen van het project doen - <!--- @huboard:{"order":105.0} -->
process
als afstudeerder wil ik een blogpost over de interessante ontdekkingen van het project doen huboard order
1
10,806
13,609,288,348
IssuesEvent
2020-09-23 04:50:13
googleapis/java-conformance-tests
https://api.github.com/repos/googleapis/java-conformance-tests
closed
Dependency Dashboard
external type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->build(deps): update dependency com.coveo:fmt-maven-plugin to v2.10 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bom-0.x -->deps: update dependency com.google.cloud:google-cloud-bom to v0.135.0 - [ ] <!-- rebase-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->deps: update dependency com.google.googlejavaformat:google-java-format to v1.9 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->build(deps): update dependency com.coveo:fmt-maven-plugin to v2.10 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bom-0.x -->deps: update dependency com.google.cloud:google-cloud-bom to v0.135.0 - [ ] <!-- rebase-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->deps: update dependency com.google.googlejavaformat:google-java-format to v1.9 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency com coveo fmt maven plugin to deps update dependency com google cloud google cloud bom to deps update dependency com google googlejavaformat google java format to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
1
8,214
2,611,471,321
IssuesEvent
2015-02-27 05:16:05
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Gentoo hedgewars crash, probably related to -O3 in system libs, causes hwengine crash possibly in sound initialization
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Start or start replaying a game with in-game sound turned on. What is the expected output? What do you see instead? Using a recorded demo and hwengine (to get the most debugging info) I run hwengine /usr/local/share/hedgewars/Data/ .hedgewars/Demos/2011-05-31_20-43.37.hwd --set-audio 100 0 1 --set-video 640 480 24 and I get the following output: Hedgewars 0.9.15 engine (network protocol: 37) Init SDL... ok Init SDL_ttf... ok Init SDL_image... ok Loading /usr/local/share/hedgewars/Data//Graphics/hwengine.png [flags: 8] ok (32x32) Loading progress sprite: Loading /usr/local/share/hedgewars/Data//Graphics/Progress.png [flags: 6] ok (324x972) Number of game controllers: 0 Not using any game controller Getting game config... Init sound...An unhandled exception occurred at $B7484F5A : EAccessViolation : Access violation $B7484F5A Without sound I run hwengine /usr/local/share/hedgewars/Data/ .hedgewars/Demos/2011-05-31_20-43.37.hwd --set-audio 100 0 0 --set-video 640 480 24 and I get (the game replay starts normally) Hedgewars 0.9.15 engine (network protocol: 37) Init SDL... ok Init SDL_ttf... ok Init SDL_image... ok Loading /usr/local/share/hedgewars/Data//Graphics/hwengine.png [flags: 8] ok (32x32) Loading progress sprite: Loading /usr/local/share/hedgewars/Data//Graphics/Progress.png [flags: 6] ok (324x972) Number of game controllers: 0 Not using any game controller Getting game config... Reading objects info... etc. What version of the product are you using? On what operating system? Hedgewars 0.9.15 on Gentoo Linux (not compiled using the Hedgewars ebuild but using the ordinary cmake/make scripts only). Please provide any additional information below. Additional system information (partly Gentoo specific): System uname: Linux-2.6.36-gentoo-i686-Intel-R-_Pentium-R-_Dual_CPU_T2370_@_1.73GHz-with-gento o-2.0.2 Timestamp of tree: Sun, 22 May 2011 07:15:01 +0000 app-shells/bash: 4.2_p10 dev-java/java-config: 2.1.11-r3 dev-lang/python: 2.7.1-r1, 3.2 dev-util/cmake: 2.8.4-r1 sys-apps/baselayout: 2.0.2 sys-apps/openrc: 0.8.2-r1 sys-apps/sandbox: 2.5 sys-devel/autoconf: 2.13, 2.68 sys-devel/automake: 1.10.3, 1.11.1-r1 sys-devel/binutils: 2.21 sys-devel/gcc: 4.5.2 sys-devel/gcc-config: 1.4.1-r1 sys-devel/libtool: 2.4-r1 sys-devel/make: 3.82 sys-kernel/linux-headers: 2.6.38 (virtual/os-headers) sys-libs/glibc: 2.13-r2 These are the sound modules loaded: $ lsmod | grep snd snd_seq_oss 23626 0 snd_seq_midi_event 4320 1 snd_seq_oss snd_seq 39947 4 snd_seq_oss,snd_seq_midi_event snd_seq_device 4149 2 snd_seq_oss,snd_seq snd_pcm_oss 32477 0 snd_mixer_oss 12525 1 snd_pcm_oss snd_hda_codec_realtek 196303 1 snd_hda_intel 17533 0 snd_hda_codec 54466 2 snd_hda_codec_realtek,snd_hda_intel snd_pcm 56347 3 snd_pcm_oss,snd_hda_intel,snd_hda_codec snd_timer 14943 2 snd_seq,snd_pcm snd 39102 10 snd_seq_oss,snd_seq,snd_seq_device,snd_pcm_oss,snd_mixer_oss,snd_hda_codec_realt ek,snd_hda_intel,snd_hda_codec,snd_pcm,snd_timer soundcore 4079 1 snd snd_page_alloc 5485 2 snd_hda_intel,snd_pcm These are the versions of SDL and SDL-Mixer I'm using: media-libs/libsdl-1.2.14-r6 USE="X alsa audio joystick opengl video xv -aalib -custom-cflags -dga -directfb -fbcon -ggi -libcaca -nas -oss (-ps3) -pulseaudio -static-libs -svga -tslib -xinerama" media-libs/sdl-mixer-1.2.11-r1 USE="flac mad midi mikmod mp3 timidity vorbis wav -static-libs" I get sound output from pretty much everything besides Hedgewars. ``` Original issue reported on code.google.com by `psyill....@gmail.com` on 31 May 2011 at 8:31
1.0
Gentoo hedgewars crash, probably related to -O3 in system libs, causes hwengine crash possibly in sound initialization - ``` What steps will reproduce the problem? 1. Start or start replaying a game with in-game sound turned on. What is the expected output? What do you see instead? Using a recorded demo and hwengine (to get the most debugging info) I run hwengine /usr/local/share/hedgewars/Data/ .hedgewars/Demos/2011-05-31_20-43.37.hwd --set-audio 100 0 1 --set-video 640 480 24 and I get the following output: Hedgewars 0.9.15 engine (network protocol: 37) Init SDL... ok Init SDL_ttf... ok Init SDL_image... ok Loading /usr/local/share/hedgewars/Data//Graphics/hwengine.png [flags: 8] ok (32x32) Loading progress sprite: Loading /usr/local/share/hedgewars/Data//Graphics/Progress.png [flags: 6] ok (324x972) Number of game controllers: 0 Not using any game controller Getting game config... Init sound...An unhandled exception occurred at $B7484F5A : EAccessViolation : Access violation $B7484F5A Without sound I run hwengine /usr/local/share/hedgewars/Data/ .hedgewars/Demos/2011-05-31_20-43.37.hwd --set-audio 100 0 0 --set-video 640 480 24 and I get (the game replay starts normally) Hedgewars 0.9.15 engine (network protocol: 37) Init SDL... ok Init SDL_ttf... ok Init SDL_image... ok Loading /usr/local/share/hedgewars/Data//Graphics/hwengine.png [flags: 8] ok (32x32) Loading progress sprite: Loading /usr/local/share/hedgewars/Data//Graphics/Progress.png [flags: 6] ok (324x972) Number of game controllers: 0 Not using any game controller Getting game config... Reading objects info... etc. What version of the product are you using? On what operating system? Hedgewars 0.9.15 on Gentoo Linux (not compiled using the Hedgewars ebuild but using the ordinary cmake/make scripts only). Please provide any additional information below. Additional system information (partly Gentoo specific): System uname: Linux-2.6.36-gentoo-i686-Intel-R-_Pentium-R-_Dual_CPU_T2370_@_1.73GHz-with-gento o-2.0.2 Timestamp of tree: Sun, 22 May 2011 07:15:01 +0000 app-shells/bash: 4.2_p10 dev-java/java-config: 2.1.11-r3 dev-lang/python: 2.7.1-r1, 3.2 dev-util/cmake: 2.8.4-r1 sys-apps/baselayout: 2.0.2 sys-apps/openrc: 0.8.2-r1 sys-apps/sandbox: 2.5 sys-devel/autoconf: 2.13, 2.68 sys-devel/automake: 1.10.3, 1.11.1-r1 sys-devel/binutils: 2.21 sys-devel/gcc: 4.5.2 sys-devel/gcc-config: 1.4.1-r1 sys-devel/libtool: 2.4-r1 sys-devel/make: 3.82 sys-kernel/linux-headers: 2.6.38 (virtual/os-headers) sys-libs/glibc: 2.13-r2 These are the sound modules loaded: $ lsmod | grep snd snd_seq_oss 23626 0 snd_seq_midi_event 4320 1 snd_seq_oss snd_seq 39947 4 snd_seq_oss,snd_seq_midi_event snd_seq_device 4149 2 snd_seq_oss,snd_seq snd_pcm_oss 32477 0 snd_mixer_oss 12525 1 snd_pcm_oss snd_hda_codec_realtek 196303 1 snd_hda_intel 17533 0 snd_hda_codec 54466 2 snd_hda_codec_realtek,snd_hda_intel snd_pcm 56347 3 snd_pcm_oss,snd_hda_intel,snd_hda_codec snd_timer 14943 2 snd_seq,snd_pcm snd 39102 10 snd_seq_oss,snd_seq,snd_seq_device,snd_pcm_oss,snd_mixer_oss,snd_hda_codec_realt ek,snd_hda_intel,snd_hda_codec,snd_pcm,snd_timer soundcore 4079 1 snd snd_page_alloc 5485 2 snd_hda_intel,snd_pcm These are the versions of SDL and SDL-Mixer I'm using: media-libs/libsdl-1.2.14-r6 USE="X alsa audio joystick opengl video xv -aalib -custom-cflags -dga -directfb -fbcon -ggi -libcaca -nas -oss (-ps3) -pulseaudio -static-libs -svga -tslib -xinerama" media-libs/sdl-mixer-1.2.11-r1 USE="flac mad midi mikmod mp3 timidity vorbis wav -static-libs" I get sound output from pretty much everything besides Hedgewars. ``` Original issue reported on code.google.com by `psyill....@gmail.com` on 31 May 2011 at 8:31
non_process
gentoo hedgewars crash probably related to in system libs causes hwengine crash possibly in sound initialization what steps will reproduce the problem start or start replaying a game with in game sound turned on what is the expected output what do you see instead using a recorded demo and hwengine to get the most debugging info i run hwengine usr local share hedgewars data hedgewars demos hwd set audio set video and i get the following output hedgewars engine network protocol init sdl ok init sdl ttf ok init sdl image ok loading usr local share hedgewars data graphics hwengine png ok loading progress sprite loading usr local share hedgewars data graphics progress png ok number of game controllers not using any game controller getting game config init sound an unhandled exception occurred at eaccessviolation access violation without sound i run hwengine usr local share hedgewars data hedgewars demos hwd set audio set video and i get the game replay starts normally hedgewars engine network protocol init sdl ok init sdl ttf ok init sdl image ok loading usr local share hedgewars data graphics hwengine png ok loading progress sprite loading usr local share hedgewars data graphics progress png ok number of game controllers not using any game controller getting game config reading objects info etc what version of the product are you using on what operating system hedgewars on gentoo linux not compiled using the hedgewars ebuild but using the ordinary cmake make scripts only please provide any additional information below additional system information partly gentoo specific system uname linux gentoo intel r pentium r dual cpu with gento o timestamp of tree sun may app shells bash dev java java config dev lang python dev util cmake sys apps baselayout sys apps openrc sys apps sandbox sys devel autoconf sys devel automake sys devel binutils sys devel gcc sys devel gcc config sys devel libtool sys devel make sys kernel linux headers virtual os headers sys libs glibc these are the sound modules loaded lsmod grep snd snd seq oss snd seq midi event snd seq oss snd seq snd seq oss snd seq midi event snd seq device snd seq oss snd seq snd pcm oss snd mixer oss snd pcm oss snd hda codec realtek snd hda intel snd hda codec snd hda codec realtek snd hda intel snd pcm snd pcm oss snd hda intel snd hda codec snd timer snd seq snd pcm snd snd seq oss snd seq snd seq device snd pcm oss snd mixer oss snd hda codec realt ek snd hda intel snd hda codec snd pcm snd timer soundcore snd snd page alloc snd hda intel snd pcm these are the versions of sdl and sdl mixer i m using media libs libsdl use x alsa audio joystick opengl video xv aalib custom cflags dga directfb fbcon ggi libcaca nas oss pulseaudio static libs svga tslib xinerama media libs sdl mixer use flac mad midi mikmod timidity vorbis wav static libs i get sound output from pretty much everything besides hedgewars original issue reported on code google com by psyill gmail com on may at
0
6,054
8,875,387,629
IssuesEvent
2019-01-12 03:02:50
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
[arm32/Windows] System.Diagnostics.Process.Tests failures on arm32 windows
arch-arm32 area-System.Diagnostics.Process os-windows-iot test-run-core
Running the arm32 tests on windows for System.Diagnostics.Process.Tests produced the following results: passed 257/261 tests. 4 tests failed the run for different reasons: System.Diagnostics.Tests.ProcessStartInfoTests.Verbs_GetWithExeExtension_ReturnsExpected ``` Assert.Contains() Failure\r\nNot found: runas\r\nIn value: String[] [\"open\"] at System.Diagnostics.Tests.ProcessStartInfoTests.Verbs_GetWithExeExtension_ReturnsExpected() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 534 ``` System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadVerb(useShellExecute: True) and System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(useShellExecute: True) ``` Assert.Throws() Failure\r\nExpected: typeof(System.ComponentModel.Win32Exception)\r\nActual: typeof(System.PlatformNotSupportedException): UseShellExecute is not supported on this platform.\r\n---- System.PlatformNotSupportedException : UseShellExecute is not supported on this platform. at System.Diagnostics.Process.StartWithShellExecuteEx(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 113 at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 24 at System.Diagnostics.Process.Start() in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1216 at System.Diagnostics.Process.Start(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1259 at System.Diagnostics.Tests.ProcessStartInfoTests.<>c__DisplayClass50_0.<StartInfo_BadVerb>b__0() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1140 ----- Inner Stack Trace ----- at System.Diagnostics.Process.StartWithShellExecuteEx(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 113 at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 24 at System.Diagnostics.Process.Start() in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1216 at System.Diagnostics.Process.Start(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1259 at System.Diagnostics.Tests.ProcessStartInfoTests.<>c__DisplayClass50_0.<StartInfo_BadVerb>b__0() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1140 ``` System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(useShellExecute: False) ``` Assert.Equal() Failure\r\nExpected: 193\r\nActual: 120 at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(Boolean useShellExecute) in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1163 ``` [testResults.zip](https://github.com/dotnet/corefx/files/2669898/testResults.zip)
1.0
[arm32/Windows] System.Diagnostics.Process.Tests failures on arm32 windows - Running the arm32 tests on windows for System.Diagnostics.Process.Tests produced the following results: passed 257/261 tests. 4 tests failed the run for different reasons: System.Diagnostics.Tests.ProcessStartInfoTests.Verbs_GetWithExeExtension_ReturnsExpected ``` Assert.Contains() Failure\r\nNot found: runas\r\nIn value: String[] [\"open\"] at System.Diagnostics.Tests.ProcessStartInfoTests.Verbs_GetWithExeExtension_ReturnsExpected() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 534 ``` System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadVerb(useShellExecute: True) and System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(useShellExecute: True) ``` Assert.Throws() Failure\r\nExpected: typeof(System.ComponentModel.Win32Exception)\r\nActual: typeof(System.PlatformNotSupportedException): UseShellExecute is not supported on this platform.\r\n---- System.PlatformNotSupportedException : UseShellExecute is not supported on this platform. at System.Diagnostics.Process.StartWithShellExecuteEx(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 113 at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 24 at System.Diagnostics.Process.Start() in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1216 at System.Diagnostics.Process.Start(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1259 at System.Diagnostics.Tests.ProcessStartInfoTests.<>c__DisplayClass50_0.<StartInfo_BadVerb>b__0() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1140 ----- Inner Stack Trace ----- at System.Diagnostics.Process.StartWithShellExecuteEx(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 113 at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.Win32.cs:line 24 at System.Diagnostics.Process.Start() in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1216 at System.Diagnostics.Process.Start(ProcessStartInfo startInfo) in F:\git\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs:line 1259 at System.Diagnostics.Tests.ProcessStartInfoTests.<>c__DisplayClass50_0.<StartInfo_BadVerb>b__0() in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1140 ``` System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(useShellExecute: False) ``` Assert.Equal() Failure\r\nExpected: 193\r\nActual: 120 at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_BadExe(Boolean useShellExecute) in F:\git\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 1163 ``` [testResults.zip](https://github.com/dotnet/corefx/files/2669898/testResults.zip)
process
system diagnostics process tests failures on windows running the tests on windows for system diagnostics process tests produced the following results passed tests tests failed the run for different reasons system diagnostics tests processstartinfotests verbs getwithexeextension returnsexpected assert contains failure r nnot found runas r nin value string at system diagnostics tests processstartinfotests verbs getwithexeextension returnsexpected in f git corefx src system diagnostics process tests processstartinfotests cs line system diagnostics tests processstartinfotests startinfo badverb useshellexecute true and system diagnostics tests processstartinfotests startinfo badexe useshellexecute true assert throws failure r nexpected typeof system componentmodel r nactual typeof system platformnotsupportedexception useshellexecute is not supported on this platform r n system platformnotsupportedexception useshellexecute is not supported on this platform at system diagnostics process startwithshellexecuteex processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process startcore processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process start in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process start processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics tests processstartinfotests c b in f git corefx src system diagnostics process tests processstartinfotests cs line inner stack trace at system diagnostics process startwithshellexecuteex processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process startcore processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process start in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics process start processstartinfo startinfo in f git corefx src system diagnostics process src system diagnostics process cs line at system diagnostics tests processstartinfotests c b in f git corefx src system diagnostics process tests processstartinfotests cs line system diagnostics tests processstartinfotests startinfo badexe useshellexecute false assert equal failure r nexpected r nactual at system diagnostics tests processstartinfotests startinfo badexe boolean useshellexecute in f git corefx src system diagnostics process tests processstartinfotests cs line
1
3,606
6,649,255,675
IssuesEvent
2017-09-28 12:36:17
usgs-bcb/bcb-dm
https://api.github.com/repos/usgs-bcb/bcb-dm
opened
Test and record CRS
DMP Acquisition DMP Distribution DMP Preservation DMP Processing DMP Product Info Protected Areas
The issue of managing the CRS came up briefly in a discussion between @dwief-usgs and @McKerrow. As we move towards more analytical and reusable code, we should ensure we test for and record the CRS of incoming spatial data. This should be a requirement for the SFR. The same mindset should be applied throughout our analytical codebase as we mature the system.
1.0
Test and record CRS - The issue of managing the CRS came up briefly in a discussion between @dwief-usgs and @McKerrow. As we move towards more analytical and reusable code, we should ensure we test for and record the CRS of incoming spatial data. This should be a requirement for the SFR. The same mindset should be applied throughout our analytical codebase as we mature the system.
process
test and record crs the issue of managing the crs came up briefly in a discussion between dwief usgs and mckerrow as we move towards more analytical and reusable code we should ensure we test for and record the crs of incoming spatial data this should be a requirement for the sfr the same mindset should be applied throughout our analytical codebase as we mature the system
1
17,214
22,821,630,576
IssuesEvent
2022-07-12 03:19:57
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
This feature "Generate XYZ tiles" display different levels of tiles
Feedback stale Raster Processing Bug
### What is the bug or the crash? This feature has bugs,toolbox->Raster tools->Generate XYZ tiles ,display different levels of tiles ![image](https://user-images.githubusercontent.com/9930104/161242493-9d9c7677-f703-448c-a890-f4d8c4885d77.png) ![image](https://user-images.githubusercontent.com/9930104/161242256-2770ff18-79c3-4e38-ab3c-698c76293e14.png) ### Steps to reproduce the issue ![image](https://user-images.githubusercontent.com/9930104/161242764-b3e8a882-78e5-4fc2-9da2-c75c097196a0.png) ### Versions v3.24.1 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
This feature "Generate XYZ tiles" display different levels of tiles - ### What is the bug or the crash? This feature has bugs,toolbox->Raster tools->Generate XYZ tiles ,display different levels of tiles ![image](https://user-images.githubusercontent.com/9930104/161242493-9d9c7677-f703-448c-a890-f4d8c4885d77.png) ![image](https://user-images.githubusercontent.com/9930104/161242256-2770ff18-79c3-4e38-ab3c-698c76293e14.png) ### Steps to reproduce the issue ![image](https://user-images.githubusercontent.com/9930104/161242764-b3e8a882-78e5-4fc2-9da2-c75c097196a0.png) ### Versions v3.24.1 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
this feature generate xyz tiles display different levels of tiles what is the bug or the crash this feature has bugs toolbox raster tools generate xyz tiles display different levels of tiles steps to reproduce the issue versions supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
412,118
12,035,524,827
IssuesEvent
2020-04-13 18:02:42
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
opened
[platform] expired transaction and conflict metrics shows incorrect values
area/platform kind/bug priority/high
Show correct values with every refresh for Expired Transactions and Conflicts in the Transactions metric shown in the DocDB section.
1.0
[platform] expired transaction and conflict metrics shows incorrect values - Show correct values with every refresh for Expired Transactions and Conflicts in the Transactions metric shown in the DocDB section.
non_process
expired transaction and conflict metrics shows incorrect values show correct values with every refresh for expired transactions and conflicts in the transactions metric shown in the docdb section
0
16,031
20,188,245,107
IssuesEvent
2022-02-11 01:21:14
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Establish process and tools to manage privileged access with just-in-time capabilities
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Separation of duties
<a href="https://docs.microsoft.com/azure/security/fundamentals/identity-management-best-practices?bc=%2fazure%2farchitecture%2fbread%2ftoc.json&toc=%2fazure%2farchitecture%2ftoc.json#lower-exposure-of-privileged-accounts">Establish process and tools to manage privileged access with just-in-time capabilities</a> <p><b>Why Consider This?</b></p> Microsoft recommends starting from the Core Services reference permissions model to define management roles and scope, and the Segment reference permissions model for workloads or landing zones, to provide clear and consistent guidance for technical teams implementing these permissions. <p><b>Context</b></p> <p><span>This clarity makes it easier to detect and correct human errors like over-permissioning, reducing organizational overall risk.</span></p> <p><b>Suggested Actions</b></p> <p><span>Implement and document a privileged access strategy sourced from Microsoft Core Services reference models.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#assign-privileges-for-managing-the-environment" target="_blank"><span>Assign privileges for managing the environment</span></a><span /></p>
1.0
Establish process and tools to manage privileged access with just-in-time capabilities - <a href="https://docs.microsoft.com/azure/security/fundamentals/identity-management-best-practices?bc=%2fazure%2farchitecture%2fbread%2ftoc.json&toc=%2fazure%2farchitecture%2ftoc.json#lower-exposure-of-privileged-accounts">Establish process and tools to manage privileged access with just-in-time capabilities</a> <p><b>Why Consider This?</b></p> Microsoft recommends starting from the Core Services reference permissions model to define management roles and scope, and the Segment reference permissions model for workloads or landing zones, to provide clear and consistent guidance for technical teams implementing these permissions. <p><b>Context</b></p> <p><span>This clarity makes it easier to detect and correct human errors like over-permissioning, reducing organizational overall risk.</span></p> <p><b>Suggested Actions</b></p> <p><span>Implement and document a privileged access strategy sourced from Microsoft Core Services reference models.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#assign-privileges-for-managing-the-environment" target="_blank"><span>Assign privileges for managing the environment</span></a><span /></p>
process
establish process and tools to manage privileged access with just in time capabilities why consider this microsoft recommends starting from the core services reference permissions model to define management roles and scope and the segment reference permissions model for workloads or landing zones to provide clear and consistent guidance for technical teams implementing these permissions context this clarity makes it easier to detect and correct human errors like over permissioning reducing organizational overall risk suggested actions implement and document a privileged access strategy sourced from microsoft core services reference models learn more assign privileges for managing the environment
1
247,790
18,857,980,768
IssuesEvent
2021-11-12 09:14:59
EthanWong6362/pe
https://api.github.com/repos/EthanWong6362/pe
opened
Reference to AddressBook instead of TaskMaster2103
severity.VeryLow type.DocumentationBug
![image.png](https://raw.githubusercontent.com/EthanWong6362/pe/main/files/483c8dc9-b93d-4956-85aa-b32dc0b23bb2.png) This link in Setting Up and Getting Started should be titled TaskMaster2103's architecture instead. <!--session: 1636703591630-5c1d2499-1a9e-4ca0-949d-0b69fb026b33--> <!--Version: Web v3.4.1-->
1.0
Reference to AddressBook instead of TaskMaster2103 - ![image.png](https://raw.githubusercontent.com/EthanWong6362/pe/main/files/483c8dc9-b93d-4956-85aa-b32dc0b23bb2.png) This link in Setting Up and Getting Started should be titled TaskMaster2103's architecture instead. <!--session: 1636703591630-5c1d2499-1a9e-4ca0-949d-0b69fb026b33--> <!--Version: Web v3.4.1-->
non_process
reference to addressbook instead of this link in setting up and getting started should be titled s architecture instead
0