Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
13,612
| 16,194,434,192
|
IssuesEvent
|
2021-05-04 13:01:01
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] UI issues with appearance of placeholder text in empty tables
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
For all PM screens that have tables and display the placeholder
1.Reduce the font size of the text
2. Reduce the darkness of the color to a faded out, lighter grey so it looks like a placeholder rather than primary text
3. Remove the full stop at the end

|
3.0
|
[PM] UI issues with appearance of placeholder text in empty tables - For all PM screens that have tables and display the placeholder
1.Reduce the font size of the text
2. Reduce the darkness of the color to a faded out, lighter grey so it looks like a placeholder rather than primary text
3. Remove the full stop at the end

|
process
|
ui issues with appearance of placeholder text in empty tables for all pm screens that have tables and display the placeholder reduce the font size of the text reduce the darkness of the color to a faded out lighter grey so it looks like a placeholder rather than primary text remove the full stop at the end
| 1
|
548,696
| 16,074,130,822
|
IssuesEvent
|
2021-04-25 02:33:53
|
fog/fog-google
|
https://api.github.com/repos/fog/fog-google
|
closed
|
Implement Licences API
|
enhancement hacktoberfest help wanted no-issue-activity priority/low
|
https://cloud.google.com/compute/docs/reference/latest/licenses
Should be quite straightfoward:
- Implement new model + collection
- Update linked resources with fields (`images` already updated)
Not sure if array inside the `images` model should be a collection of `License` objects or not though.
|
1.0
|
Implement Licences API - https://cloud.google.com/compute/docs/reference/latest/licenses
Should be quite straightfoward:
- Implement new model + collection
- Update linked resources with fields (`images` already updated)
Not sure if array inside the `images` model should be a collection of `License` objects or not though.
|
non_process
|
implement licences api should be quite straightfoward implement new model collection update linked resources with fields images already updated not sure if array inside the images model should be a collection of license objects or not though
| 0
|
22,731
| 32,049,835,021
|
IssuesEvent
|
2023-09-23 12:25:45
|
plazi/monitoring
|
https://api.github.com/repos/plazi/monitoring
|
opened
|
🛑 XML Processing Actions is down
|
status xml-processing-actions
|
In [`58ba03e`](https://github.com/plazi/monitoring/commit/58ba03e467ada9e74fbc4c494678f4f7c2448483
), XML Processing Actions (https://github.com/plazi/treatments-xml/actions/workflows/main.yml/badge.svg) was **down**:
- HTTP code: 200
- Response time: 331 ms
|
1.0
|
🛑 XML Processing Actions is down - In [`58ba03e`](https://github.com/plazi/monitoring/commit/58ba03e467ada9e74fbc4c494678f4f7c2448483
), XML Processing Actions (https://github.com/plazi/treatments-xml/actions/workflows/main.yml/badge.svg) was **down**:
- HTTP code: 200
- Response time: 331 ms
|
process
|
🛑 xml processing actions is down in xml processing actions was down http code response time ms
| 1
|
235,986
| 19,477,272,438
|
IssuesEvent
|
2021-12-24 15:20:56
|
eclipse/che
|
https://api.github.com/repos/eclipse/che
|
closed
|
"chectl server:deploy" command failed because of error "open /tmp/keycloak-provision.sh: no such file or directory"
|
kind/bug e2e-test/failure area/chectl area/operator
|
### Describe the bug
https://main-jenkins-csb-crwqe.apps.ocp4.prod.psi.redhat.com/job/Che/job/e2e/job/minikube/job/basic/job/che-server/1364/console
```
(node:6764) Warning: Setting the NODE_TLS_REJECT_UNAUTHORIZED environment variable to '0' makes TLS connections and HTTPS requests insecure by disabling certificate verification.
› Installer type is set to: 'operator'
[07:38:28] Verify Kubernetes API [started]
› Current Kubernetes context: 'minikube'
[07:38:28] Verify Kubernetes API...OK [title changed]
[07:38:28] Verify Kubernetes API...OK [completed]
[07:38:28] 👀 Looking for an already existing Eclipse Che instance [started]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che" [started]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che"...it is not [title changed]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che"...it is not [completed]
[07:38:28] 👀 Looking for an already existing Eclipse Che instance [completed]
[07:38:28] ✈️ Minikube preflight checklist [started]
[07:38:28] Verify if kubectl is installed [started]
[07:38:28] Verify if kubectl is installed [completed]
[07:38:28] Verify if minikube is installed [started]
[07:38:28] Verify if minikube is installed [completed]
[07:38:28] Verify if minikube is running [started]
[07:38:29] Verify if minikube is running [completed]
[07:38:29] Start minikube [started]
[07:38:29] Start minikube [skipped]
[07:38:29] → Minikube is already running.
[07:38:29] Check Kubernetes version [started]
[07:38:29] Check Kubernetes version: Found v1.20.2. [title changed]
[07:38:29] Check Kubernetes version: Found v1.20.2. [completed]
[07:38:29] Verify if minikube ingress addon is enabled [started]
[07:38:29] Verify if minikube ingress addon is enabled [completed]
[07:38:29] Enable minikube ingress addon [started]
[07:38:40] Enable minikube ingress addon [completed]
[07:38:40] Retrieving minikube IP and domain for ingress URLs [started]
[07:38:41] Retrieving minikube IP and domain for ingress URLs...10.0.209.178.nip.io. [title changed]
[07:38:41] Retrieving minikube IP and domain for ingress URLs...10.0.209.178.nip.io. [completed]
[07:38:41] Checking minikube version [started]
[07:38:41] Checking minikube version... 1.22.0 [title changed]
[07:38:41] Checking minikube version... 1.22.0 [completed]
[07:38:41] Check if cluster accessible [started]
[07:38:41] Check if cluster accessible [skipped]
[07:38:41] ✈️ Minikube preflight checklist [completed]
[07:38:41] Following Eclipse Che logs [started]
[07:38:41] Start following Operator logs [started]
[07:38:41] Start following Operator logs...done [title changed]
[07:38:41] Start following Operator logs...done [completed]
[07:38:41] Start following Eclipse Che Server logs [started]
[07:38:41] Start following Eclipse Che Server logs...done [title changed]
[07:38:41] Start following Eclipse Che Server logs...done [completed]
[07:38:41] Start following PostgreSQL logs [started]
[07:38:41] Start following PostgreSQL logs...done [title changed]
[07:38:41] Start following PostgreSQL logs...done [completed]
[07:38:41] Start following Keycloak logs [started]
[07:38:41] Start following Keycloak logs...done [title changed]
[07:38:41] Start following Keycloak logs...done [completed]
[07:38:41] Start following Plug-in Registry logs [started]
[07:38:41] Start following Plug-in Registry logs...done [title changed]
[07:38:41] Start following Plug-in Registry logs...done [completed]
[07:38:41] Start following Devfile Registry logs [started]
[07:38:41] Start following Devfile Registry logs...done [title changed]
[07:38:41] Start following Devfile Registry logs...done [completed]
[07:38:41] Start following Eclipse Che Dashboard logs [started]
[07:38:41] Start following Eclipse Che Dashboard logs...done [title changed]
[07:38:41] Start following Eclipse Che Dashboard logs...done [completed]
[07:38:41] Start following namespace events [started]
[07:38:41] Start following namespace events...done [title changed]
[07:38:41] Start following namespace events...done [completed]
[07:38:41] Following Eclipse Che logs [completed]
[07:38:41] Create Namespace eclipse-che [started]
[07:38:41] Create Namespace eclipse-che...[OK] [title changed]
[07:38:41] Create Namespace eclipse-che...[OK] [completed]
[07:38:41] 🏃 Running the Eclipse Che operator [started]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che [started]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che...done. [title changed]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che...done. [completed]
[07:38:41] Read Roles and Bindings [started]
[07:38:41] Read Roles and Bindings...done. [title changed]
[07:38:41] Read Roles and Bindings...done. [completed]
[07:38:41] Creating Roles and Bindings [started]
[07:38:42] Creating Roles and Bindings...done. [title changed]
[07:38:42] Creating Roles and Bindings...done. [completed]
[07:38:42] Create CRD checlusters.org.eclipse.che [started]
[07:38:42] Create CRD checlusters.org.eclipse.che...done. [title changed]
[07:38:42] Create CRD checlusters.org.eclipse.che...done. [completed]
[07:38:42] Create backup and restore CRDs [started]
[07:38:43] Create backup and restore CRDs...done. [title changed]
[07:38:43] Create backup and restore CRDs...done. [completed]
[07:38:43] Waiting 5 seconds for the new Kubernetes resources to get flushed [started]
[07:38:48] Waiting 5 seconds for the new Kubernetes resources to get flushed...done. [title changed]
[07:38:48] Waiting 5 seconds for the new Kubernetes resources to get flushed...done. [completed]
[07:38:48] Create deployment che-operator in namespace eclipse-che [started]
[07:38:48] Create deployment che-operator in namespace eclipse-che...done. [title changed]
[07:38:48] Create deployment che-operator in namespace eclipse-che...done. [completed]
[07:38:48] Operator pod bootstrap [started]
[07:38:48] Scheduling [started]
[07:38:49] Scheduling...done [title changed]
[07:38:49] Scheduling...done [completed]
[07:38:49] Downloading images [started]
[07:39:07] Downloading images...done [title changed]
[07:39:07] Downloading images...done [completed]
[07:39:07] Starting [started]
[07:39:13] Starting...done [title changed]
[07:39:13] Starting...done [completed]
[07:39:13] Operator pod bootstrap [completed]
[07:39:13] Prepare Eclipse Che cluster CR [started]
[07:39:13] Prepare Eclipse Che cluster CR...Done. [title changed]
[07:39:13] Prepare Eclipse Che cluster CR...Done. [completed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che [started]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che [title changed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che...done. [title changed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che...done. [completed]
[07:39:13] 🏃 Running the Eclipse Che operator [completed]
[07:39:13] ✅ Post installation checklist [started]
[07:39:13] PostgreSQL pod bootstrap [started]
[07:39:13] Scheduling [started]
[07:39:33] Scheduling...done [title changed]
[07:39:33] Scheduling...done [completed]
[07:39:33] Downloading images [started]
[07:40:12] Downloading images...done [title changed]
[07:40:12] Downloading images...done [completed]
[07:40:12] Starting [started]
[07:40:31] Starting...done [title changed]
[07:40:31] Starting...done [completed]
[07:40:31] PostgreSQL pod bootstrap [completed]
[07:40:31] Keycloak pod bootstrap [started]
[07:40:31] Scheduling [started]
[07:40:33] Scheduling...done [title changed]
[07:40:33] Scheduling...done [completed]
[07:40:33] Downloading images [started]
[07:41:13] Downloading images...done [title changed]
[07:41:13] Downloading images...done [completed]
[07:41:13] Starting [started]
[07:42:44] Starting...done [title changed]
[07:42:44] Starting...done [completed]
[07:42:44] Keycloak pod bootstrap [completed]
[07:42:44] Devfile Registry pod bootstrap [started]
[07:42:44] Scheduling [started]
[07:45:04] Scheduling...failed [title changed]
[07:45:04] Scheduling...failed [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
[07:45:04] Devfile Registry pod bootstrap [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
[07:45:04] ✅ Post installation checklist [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
Error: Command server:deploy failed. Error log:
/home/hudson/.cache/chectl/error.log.
```
### Che version
next (development version)
### Steps to reproduce
1. Start minikube in 'none' driver mode
2. Download chectl next: https://github.com/che-incubator/chectl/releases/download/20211217131027/chectl-linux-x64.tar.gz
3. Run command:
```
chectl server:deploy --k8spodreadytimeout=600000 --k8spodwaittimeout=600000 --k8spoddownloadimagetimeout=600000 --listr-renderer=verbose --platform=minikube --che-operator-cr-patch-yaml=custom-resource-patch.yaml --telemetry=off --chenamespace=eclipse-che
```
where custom-resource-patch.yaml:
```
spec:
server:
cheImage: 'quay.io/eclipse/che-server'
cheImageTag: 'next'
customCheProperties:
CHE_LIMITS_USER_WORKSPACES_RUN_COUNT: '-1'
CHE_WORKSPACE_SIDECAR_IMAGE__PULL__POLICY: IfNotPresent
CHE_WORKSPACE_PLUGIN__BROKER_PULL__POLICY: IfNotPresent
CHE_INFRA_KUBERNETES_WORKSPACE_START_TIMEOUT_MIN: '12'
auth:
updateAdminPassword: false
```
### Expected behavior
Eclipse Che next to be deployed successfully.
### Runtime
minikube
### Screenshots
_No response_
### Installation method
chectl/next
### Environment
Linux
### Eclipse Che Logs
che-operator pod log https://main-jenkins-csb-crwqe.apps.ocp4.prod.psi.redhat.com/job/Che/job/e2e/job/minikube/job/basic/job/che-server/1364/artifact/logs-and-configs/che-logs/che-operator-6bf7cfb477-9zxsl.pod.log/*view*/
```
time="2021-12-24T12:42:49Z" level=info msg="Deployment keycloak is in the rolling update state."
time="2021-12-24T12:42:51Z" level=info msg="Deployment keycloak is in the rolling update state."
time="2021-12-24T12:44:01Z" level=info msg="Custom resource status eclipse-che updated with status: Reason: InstallOrUpdateFailed"
time="2021-12-24T12:44:01Z" level=info msg="Custom resource status eclipse-che updated with status: Message: open /tmp/keycloak-provision.sh: no such file or directory"
2021-12-24T12:44:01.333Z ERROR controller-runtime.manager.controller.checluster Reconciler error {"reconciler group": "org.eclipse.che", "reconciler kind": "CheCluster", "name": "eclipse-che", "namespace": "eclipse-che", "error": "open /tmp/keycloak-provision.sh: no such file or directory"}
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem
/che-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:253
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2
/che-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:214
```
### Additional context
It could be related to recent switch to DevWorkspace https://github.com/eclipse-che/che-operator/pull/1265
|
1.0
|
"chectl server:deploy" command failed because of error "open /tmp/keycloak-provision.sh: no such file or directory" - ### Describe the bug
https://main-jenkins-csb-crwqe.apps.ocp4.prod.psi.redhat.com/job/Che/job/e2e/job/minikube/job/basic/job/che-server/1364/console
```
(node:6764) Warning: Setting the NODE_TLS_REJECT_UNAUTHORIZED environment variable to '0' makes TLS connections and HTTPS requests insecure by disabling certificate verification.
› Installer type is set to: 'operator'
[07:38:28] Verify Kubernetes API [started]
› Current Kubernetes context: 'minikube'
[07:38:28] Verify Kubernetes API...OK [title changed]
[07:38:28] Verify Kubernetes API...OK [completed]
[07:38:28] 👀 Looking for an already existing Eclipse Che instance [started]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che" [started]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che"...it is not [title changed]
[07:38:28] Verify if Eclipse Che is deployed into namespace "eclipse-che"...it is not [completed]
[07:38:28] 👀 Looking for an already existing Eclipse Che instance [completed]
[07:38:28] ✈️ Minikube preflight checklist [started]
[07:38:28] Verify if kubectl is installed [started]
[07:38:28] Verify if kubectl is installed [completed]
[07:38:28] Verify if minikube is installed [started]
[07:38:28] Verify if minikube is installed [completed]
[07:38:28] Verify if minikube is running [started]
[07:38:29] Verify if minikube is running [completed]
[07:38:29] Start minikube [started]
[07:38:29] Start minikube [skipped]
[07:38:29] → Minikube is already running.
[07:38:29] Check Kubernetes version [started]
[07:38:29] Check Kubernetes version: Found v1.20.2. [title changed]
[07:38:29] Check Kubernetes version: Found v1.20.2. [completed]
[07:38:29] Verify if minikube ingress addon is enabled [started]
[07:38:29] Verify if minikube ingress addon is enabled [completed]
[07:38:29] Enable minikube ingress addon [started]
[07:38:40] Enable minikube ingress addon [completed]
[07:38:40] Retrieving minikube IP and domain for ingress URLs [started]
[07:38:41] Retrieving minikube IP and domain for ingress URLs...10.0.209.178.nip.io. [title changed]
[07:38:41] Retrieving minikube IP and domain for ingress URLs...10.0.209.178.nip.io. [completed]
[07:38:41] Checking minikube version [started]
[07:38:41] Checking minikube version... 1.22.0 [title changed]
[07:38:41] Checking minikube version... 1.22.0 [completed]
[07:38:41] Check if cluster accessible [started]
[07:38:41] Check if cluster accessible [skipped]
[07:38:41] ✈️ Minikube preflight checklist [completed]
[07:38:41] Following Eclipse Che logs [started]
[07:38:41] Start following Operator logs [started]
[07:38:41] Start following Operator logs...done [title changed]
[07:38:41] Start following Operator logs...done [completed]
[07:38:41] Start following Eclipse Che Server logs [started]
[07:38:41] Start following Eclipse Che Server logs...done [title changed]
[07:38:41] Start following Eclipse Che Server logs...done [completed]
[07:38:41] Start following PostgreSQL logs [started]
[07:38:41] Start following PostgreSQL logs...done [title changed]
[07:38:41] Start following PostgreSQL logs...done [completed]
[07:38:41] Start following Keycloak logs [started]
[07:38:41] Start following Keycloak logs...done [title changed]
[07:38:41] Start following Keycloak logs...done [completed]
[07:38:41] Start following Plug-in Registry logs [started]
[07:38:41] Start following Plug-in Registry logs...done [title changed]
[07:38:41] Start following Plug-in Registry logs...done [completed]
[07:38:41] Start following Devfile Registry logs [started]
[07:38:41] Start following Devfile Registry logs...done [title changed]
[07:38:41] Start following Devfile Registry logs...done [completed]
[07:38:41] Start following Eclipse Che Dashboard logs [started]
[07:38:41] Start following Eclipse Che Dashboard logs...done [title changed]
[07:38:41] Start following Eclipse Che Dashboard logs...done [completed]
[07:38:41] Start following namespace events [started]
[07:38:41] Start following namespace events...done [title changed]
[07:38:41] Start following namespace events...done [completed]
[07:38:41] Following Eclipse Che logs [completed]
[07:38:41] Create Namespace eclipse-che [started]
[07:38:41] Create Namespace eclipse-che...[OK] [title changed]
[07:38:41] Create Namespace eclipse-che...[OK] [completed]
[07:38:41] 🏃 Running the Eclipse Che operator [started]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che [started]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che...done. [title changed]
[07:38:41] Create ServiceAccount che-operator in namespace eclipse-che...done. [completed]
[07:38:41] Read Roles and Bindings [started]
[07:38:41] Read Roles and Bindings...done. [title changed]
[07:38:41] Read Roles and Bindings...done. [completed]
[07:38:41] Creating Roles and Bindings [started]
[07:38:42] Creating Roles and Bindings...done. [title changed]
[07:38:42] Creating Roles and Bindings...done. [completed]
[07:38:42] Create CRD checlusters.org.eclipse.che [started]
[07:38:42] Create CRD checlusters.org.eclipse.che...done. [title changed]
[07:38:42] Create CRD checlusters.org.eclipse.che...done. [completed]
[07:38:42] Create backup and restore CRDs [started]
[07:38:43] Create backup and restore CRDs...done. [title changed]
[07:38:43] Create backup and restore CRDs...done. [completed]
[07:38:43] Waiting 5 seconds for the new Kubernetes resources to get flushed [started]
[07:38:48] Waiting 5 seconds for the new Kubernetes resources to get flushed...done. [title changed]
[07:38:48] Waiting 5 seconds for the new Kubernetes resources to get flushed...done. [completed]
[07:38:48] Create deployment che-operator in namespace eclipse-che [started]
[07:38:48] Create deployment che-operator in namespace eclipse-che...done. [title changed]
[07:38:48] Create deployment che-operator in namespace eclipse-che...done. [completed]
[07:38:48] Operator pod bootstrap [started]
[07:38:48] Scheduling [started]
[07:38:49] Scheduling...done [title changed]
[07:38:49] Scheduling...done [completed]
[07:38:49] Downloading images [started]
[07:39:07] Downloading images...done [title changed]
[07:39:07] Downloading images...done [completed]
[07:39:07] Starting [started]
[07:39:13] Starting...done [title changed]
[07:39:13] Starting...done [completed]
[07:39:13] Operator pod bootstrap [completed]
[07:39:13] Prepare Eclipse Che cluster CR [started]
[07:39:13] Prepare Eclipse Che cluster CR...Done. [title changed]
[07:39:13] Prepare Eclipse Che cluster CR...Done. [completed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che [started]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che [title changed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che...done. [title changed]
[07:39:13] Create the Custom Resource of type checlusters.org.eclipse.che in the namespace eclipse-che...done. [completed]
[07:39:13] 🏃 Running the Eclipse Che operator [completed]
[07:39:13] ✅ Post installation checklist [started]
[07:39:13] PostgreSQL pod bootstrap [started]
[07:39:13] Scheduling [started]
[07:39:33] Scheduling...done [title changed]
[07:39:33] Scheduling...done [completed]
[07:39:33] Downloading images [started]
[07:40:12] Downloading images...done [title changed]
[07:40:12] Downloading images...done [completed]
[07:40:12] Starting [started]
[07:40:31] Starting...done [title changed]
[07:40:31] Starting...done [completed]
[07:40:31] PostgreSQL pod bootstrap [completed]
[07:40:31] Keycloak pod bootstrap [started]
[07:40:31] Scheduling [started]
[07:40:33] Scheduling...done [title changed]
[07:40:33] Scheduling...done [completed]
[07:40:33] Downloading images [started]
[07:41:13] Downloading images...done [title changed]
[07:41:13] Downloading images...done [completed]
[07:41:13] Starting [started]
[07:42:44] Starting...done [title changed]
[07:42:44] Starting...done [completed]
[07:42:44] Keycloak pod bootstrap [completed]
[07:42:44] Devfile Registry pod bootstrap [started]
[07:42:44] Scheduling [started]
[07:45:04] Scheduling...failed [title changed]
[07:45:04] Scheduling...failed [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
[07:45:04] Devfile Registry pod bootstrap [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
[07:45:04] ✅ Post installation checklist [failed]
[07:45:04] → Eclipse Che operator failed, reason: InstallOrUpdateFailed, message: open /tmp/keycloak-provision.sh: no such file or directory. Consider increasing error recheck timeout with --k8spoderrorrechecktimeout flag.
Error: Command server:deploy failed. Error log:
/home/hudson/.cache/chectl/error.log.
```
### Che version
next (development version)
### Steps to reproduce
1. Start minikube in 'none' driver mode
2. Download chectl next: https://github.com/che-incubator/chectl/releases/download/20211217131027/chectl-linux-x64.tar.gz
3. Run command:
```
chectl server:deploy --k8spodreadytimeout=600000 --k8spodwaittimeout=600000 --k8spoddownloadimagetimeout=600000 --listr-renderer=verbose --platform=minikube --che-operator-cr-patch-yaml=custom-resource-patch.yaml --telemetry=off --chenamespace=eclipse-che
```
where custom-resource-patch.yaml:
```
spec:
server:
cheImage: 'quay.io/eclipse/che-server'
cheImageTag: 'next'
customCheProperties:
CHE_LIMITS_USER_WORKSPACES_RUN_COUNT: '-1'
CHE_WORKSPACE_SIDECAR_IMAGE__PULL__POLICY: IfNotPresent
CHE_WORKSPACE_PLUGIN__BROKER_PULL__POLICY: IfNotPresent
CHE_INFRA_KUBERNETES_WORKSPACE_START_TIMEOUT_MIN: '12'
auth:
updateAdminPassword: false
```
### Expected behavior
Eclipse Che next to be deployed successfully.
### Runtime
minikube
### Screenshots
_No response_
### Installation method
chectl/next
### Environment
Linux
### Eclipse Che Logs
che-operator pod log https://main-jenkins-csb-crwqe.apps.ocp4.prod.psi.redhat.com/job/Che/job/e2e/job/minikube/job/basic/job/che-server/1364/artifact/logs-and-configs/che-logs/che-operator-6bf7cfb477-9zxsl.pod.log/*view*/
```
time="2021-12-24T12:42:49Z" level=info msg="Deployment keycloak is in the rolling update state."
time="2021-12-24T12:42:51Z" level=info msg="Deployment keycloak is in the rolling update state."
time="2021-12-24T12:44:01Z" level=info msg="Custom resource status eclipse-che updated with status: Reason: InstallOrUpdateFailed"
time="2021-12-24T12:44:01Z" level=info msg="Custom resource status eclipse-che updated with status: Message: open /tmp/keycloak-provision.sh: no such file or directory"
2021-12-24T12:44:01.333Z ERROR controller-runtime.manager.controller.checluster Reconciler error {"reconciler group": "org.eclipse.che", "reconciler kind": "CheCluster", "name": "eclipse-che", "namespace": "eclipse-che", "error": "open /tmp/keycloak-provision.sh: no such file or directory"}
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem
/che-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:253
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2
/che-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:214
```
### Additional context
It could be related to recent switch to DevWorkspace https://github.com/eclipse-che/che-operator/pull/1265
|
non_process
|
chectl server deploy command failed because of error open tmp keycloak provision sh no such file or directory describe the bug node warning setting the node tls reject unauthorized environment variable to makes tls connections and https requests insecure by disabling certificate verification › installer type is set to operator verify kubernetes api › current kubernetes context minikube verify kubernetes api ok verify kubernetes api ok 👀 looking for an already existing eclipse che instance verify if eclipse che is deployed into namespace eclipse che verify if eclipse che is deployed into namespace eclipse che it is not verify if eclipse che is deployed into namespace eclipse che it is not 👀 looking for an already existing eclipse che instance ✈️ minikube preflight checklist verify if kubectl is installed verify if kubectl is installed verify if minikube is installed verify if minikube is installed verify if minikube is running verify if minikube is running start minikube start minikube → minikube is already running check kubernetes version check kubernetes version found check kubernetes version found verify if minikube ingress addon is enabled verify if minikube ingress addon is enabled enable minikube ingress addon enable minikube ingress addon retrieving minikube ip and domain for ingress urls retrieving minikube ip and domain for ingress urls nip io retrieving minikube ip and domain for ingress urls nip io checking minikube version checking minikube version checking minikube version check if cluster accessible check if cluster accessible ✈️ minikube preflight checklist following eclipse che logs start following operator logs start following operator logs done start following operator logs done start following eclipse che server logs start following eclipse che server logs done start following eclipse che server logs done start following postgresql logs start following postgresql logs done start following postgresql logs done start following keycloak logs start following keycloak logs done start following keycloak logs done start following plug in registry logs start following plug in registry logs done start following plug in registry logs done start following devfile registry logs start following devfile registry logs done start following devfile registry logs done start following eclipse che dashboard logs start following eclipse che dashboard logs done start following eclipse che dashboard logs done start following namespace events start following namespace events done start following namespace events done following eclipse che logs create namespace eclipse che create namespace eclipse che create namespace eclipse che 🏃 running the eclipse che operator create serviceaccount che operator in namespace eclipse che create serviceaccount che operator in namespace eclipse che done create serviceaccount che operator in namespace eclipse che done read roles and bindings read roles and bindings done read roles and bindings done creating roles and bindings creating roles and bindings done creating roles and bindings done create crd checlusters org eclipse che create crd checlusters org eclipse che done create crd checlusters org eclipse che done create backup and restore crds create backup and restore crds done create backup and restore crds done waiting seconds for the new kubernetes resources to get flushed waiting seconds for the new kubernetes resources to get flushed done waiting seconds for the new kubernetes resources to get flushed done create deployment che operator in namespace eclipse che create deployment che operator in namespace eclipse che done create deployment che operator in namespace eclipse che done operator pod bootstrap scheduling scheduling done scheduling done downloading images downloading images done downloading images done starting starting done starting done operator pod bootstrap prepare eclipse che cluster cr prepare eclipse che cluster cr done prepare eclipse che cluster cr done create the custom resource of type checlusters org eclipse che create the custom resource of type checlusters org eclipse che in the namespace eclipse che create the custom resource of type checlusters org eclipse che in the namespace eclipse che done create the custom resource of type checlusters org eclipse che in the namespace eclipse che done 🏃 running the eclipse che operator ✅ post installation checklist postgresql pod bootstrap scheduling scheduling done scheduling done downloading images downloading images done downloading images done starting starting done starting done postgresql pod bootstrap keycloak pod bootstrap scheduling scheduling done scheduling done downloading images downloading images done downloading images done starting starting done starting done keycloak pod bootstrap devfile registry pod bootstrap scheduling scheduling failed scheduling failed → eclipse che operator failed reason installorupdatefailed message open tmp keycloak provision sh no such file or directory consider increasing error recheck timeout with flag devfile registry pod bootstrap → eclipse che operator failed reason installorupdatefailed message open tmp keycloak provision sh no such file or directory consider increasing error recheck timeout with flag ✅ post installation checklist → eclipse che operator failed reason installorupdatefailed message open tmp keycloak provision sh no such file or directory consider increasing error recheck timeout with flag error command server deploy failed error log home hudson cache chectl error log che version next development version steps to reproduce start minikube in none driver mode download chectl next run command chectl server deploy listr renderer verbose platform minikube che operator cr patch yaml custom resource patch yaml telemetry off chenamespace eclipse che where custom resource patch yaml spec server cheimage quay io eclipse che server cheimagetag next customcheproperties che limits user workspaces run count che workspace sidecar image pull policy ifnotpresent che workspace plugin broker pull policy ifnotpresent che infra kubernetes workspace start timeout min auth updateadminpassword false expected behavior eclipse che next to be deployed successfully runtime minikube screenshots no response installation method chectl next environment linux eclipse che logs che operator pod log time level info msg deployment keycloak is in the rolling update state time level info msg deployment keycloak is in the rolling update state time level info msg custom resource status eclipse che updated with status reason installorupdatefailed time level info msg custom resource status eclipse che updated with status message open tmp keycloak provision sh no such file or directory error controller runtime manager controller checluster reconciler error reconciler group org eclipse che reconciler kind checluster name eclipse che namespace eclipse che error open tmp keycloak provision sh no such file or directory sigs io controller runtime pkg internal controller controller processnextworkitem che operator vendor sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller start che operator vendor sigs io controller runtime pkg internal controller controller go additional context it could be related to recent switch to devworkspace
| 0
|
4,930
| 7,795,459,233
|
IssuesEvent
|
2018-06-08 08:11:50
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Plotgen3 does not support color arrays (Trac #148)
|
Migrated from Trac enhancement nielsenb@uwm.edu post_processing
|
Plotgen3 currently only can handle colors specified as strings MATLAB understands (ex. 'red', 'green'), it needs to be able to support RGB color arrays like the current plotgen (ex. [1.0, 0.8, 0.2]).
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/148
```json
{
"status": "closed",
"changetime": "2009-09-01T20:36:32",
"description": "Plotgen3 currently only can handle colors specified as strings MATLAB understands (ex. 'red', 'green'), it needs to be able to support RGB color arrays like the current plotgen (ex. [1.0, 0.8, 0.2]).",
"reporter": "nielsenb@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251837392000000",
"component": "post_processing",
"summary": "Plotgen3 does not support color arrays",
"priority": "major",
"keywords": "plotgen",
"time": "2009-07-29T15:58:36",
"milestone": "Plotgen 3.0",
"owner": "nielsenb@uwm.edu",
"type": "enhancement"
}
```
|
1.0
|
Plotgen3 does not support color arrays (Trac #148) - Plotgen3 currently only can handle colors specified as strings MATLAB understands (ex. 'red', 'green'), it needs to be able to support RGB color arrays like the current plotgen (ex. [1.0, 0.8, 0.2]).
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/148
```json
{
"status": "closed",
"changetime": "2009-09-01T20:36:32",
"description": "Plotgen3 currently only can handle colors specified as strings MATLAB understands (ex. 'red', 'green'), it needs to be able to support RGB color arrays like the current plotgen (ex. [1.0, 0.8, 0.2]).",
"reporter": "nielsenb@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251837392000000",
"component": "post_processing",
"summary": "Plotgen3 does not support color arrays",
"priority": "major",
"keywords": "plotgen",
"time": "2009-07-29T15:58:36",
"milestone": "Plotgen 3.0",
"owner": "nielsenb@uwm.edu",
"type": "enhancement"
}
```
|
process
|
does not support color arrays trac currently only can handle colors specified as strings matlab understands ex red green it needs to be able to support rgb color arrays like the current plotgen ex attachments migrated from json status closed changetime description currently only can handle colors specified as strings matlab understands ex red green it needs to be able to support rgb color arrays like the current plotgen ex reporter nielsenb uwm edu cc resolution verified by v larson ts component post processing summary does not support color arrays priority major keywords plotgen time milestone plotgen owner nielsenb uwm edu type enhancement
| 1
|
4,273
| 7,189,624,746
|
IssuesEvent
|
2018-02-02 14:40:37
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
A new tool called "makeClaim"
|
status-inprocess tools-makeClaim type-enhancement
|
makeClaim will take a string and post a transaction to the blockchain where that string is embedded in the 'input' data field of the transaction. It will send .01 (a finney) from an account to itself. It has to somehow use Parity's signer safely.
|
1.0
|
A new tool called "makeClaim" - makeClaim will take a string and post a transaction to the blockchain where that string is embedded in the 'input' data field of the transaction. It will send .01 (a finney) from an account to itself. It has to somehow use Parity's signer safely.
|
process
|
a new tool called makeclaim makeclaim will take a string and post a transaction to the blockchain where that string is embedded in the input data field of the transaction it will send a finney from an account to itself it has to somehow use parity s signer safely
| 1
|
98,058
| 20,605,112,711
|
IssuesEvent
|
2022-03-06 21:21:20
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[3+] com_joomlaupdate allows upload of any filetype
|
No Code Attached Yet
|
### Steps to reproduce the issue
In Joomla 3+ go to com_joomlaupdate ->Upload & Update tab
Select a PNG file, or XML file or any-other-file-type
### Expected result
Validation error - its not a zip file.
### Actual result
The file is uploaded to the `tmp path`, renamed and prefixed with `'ju`
The file extension and mime type is not validated first.
Then you get a login screen.
### Additional comments
Tested Joomla 3.9.19 and Joomla 4 beta 1
Cant think of anyway to exploit this so posting publicly.
|
1.0
|
[3+] com_joomlaupdate allows upload of any filetype - ### Steps to reproduce the issue
In Joomla 3+ go to com_joomlaupdate ->Upload & Update tab
Select a PNG file, or XML file or any-other-file-type
### Expected result
Validation error - its not a zip file.
### Actual result
The file is uploaded to the `tmp path`, renamed and prefixed with `'ju`
The file extension and mime type is not validated first.
Then you get a login screen.
### Additional comments
Tested Joomla 3.9.19 and Joomla 4 beta 1
Cant think of anyway to exploit this so posting publicly.
|
non_process
|
com joomlaupdate allows upload of any filetype steps to reproduce the issue in joomla go to com joomlaupdate upload update tab select a png file or xml file or any other file type expected result validation error its not a zip file actual result the file is uploaded to the tmp path renamed and prefixed with ju the file extension and mime type is not validated first then you get a login screen additional comments tested joomla and joomla beta cant think of anyway to exploit this so posting publicly
| 0
|
22,215
| 30,763,539,251
|
IssuesEvent
|
2023-07-30 02:00:07
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Fri, 28 Jul 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### A Memory-Augmented Multi-Task Collaborative Framework for Unsupervised Traffic Accident Detection in Driving Videos
- **Authors:** Rongqin Liang, Yuanman Li, Yingxin Yi, Jiantao Zhou, Xia Li
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.14575
- **Pdf link:** https://arxiv.org/pdf/2307.14575
- **Abstract**
Identifying traffic accidents in driving videos is crucial to ensuring the safety of autonomous driving and driver assistance systems. To address the potential danger caused by the long-tailed distribution of driving events, existing traffic accident detection (TAD) methods mainly rely on unsupervised learning. However, TAD is still challenging due to the rapid movement of cameras and dynamic scenes in driving scenarios. Existing unsupervised TAD methods mainly rely on a single pretext task, i.e., an appearance-based or future object localization task, to detect accidents. However, appearance-based approaches are easily disturbed by the rapid movement of the camera and changes in illumination, which significantly reduce the performance of traffic accident detection. Methods based on future object localization may fail to capture appearance changes in video frames, making it difficult to detect ego-involved accidents (e.g., out of control of the ego-vehicle). In this paper, we propose a novel memory-augmented multi-task collaborative framework (MAMTCF) for unsupervised traffic accident detection in driving videos. Different from previous approaches, our method can more accurately detect both ego-involved and non-ego accidents by simultaneously modeling appearance changes and object motions in video frames through the collaboration of optical flow reconstruction and future object localization tasks. Further, we introduce a memory-augmented motion representation mechanism to fully explore the interrelation between different types of motion representations and exploit the high-level features of normal traffic patterns stored in memory to augment motion representations, thus enlarging the difference from anomalies. Experimental results on recently published large-scale dataset demonstrate that our method achieves better performance compared to previous state-of-the-art approaches.
### To Adapt or Not to Adapt? Real-Time Adaptation for Semantic Segmentation
- **Authors:** Marc Botet Colomer, Pier Luigi Dovesi, Theodoros Panagiotakopoulos, Joao Frederico Carvalho, Linus Härenstam-Nielsen, Hossein Azizpour, Hedvig Kjellström, Daniel Cremers, Matteo Poggi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.15063
- **Pdf link:** https://arxiv.org/pdf/2307.15063
- **Abstract**
The goal of Online Domain Adaptation for semantic segmentation is to handle unforeseeable domain changes that occur during deployment, like sudden weather events. However, the high computational costs associated with brute-force adaptation make this paradigm unfeasible for real-world applications. In this paper we propose HAMLET, a Hardware-Aware Modular Least Expensive Training framework for real-time domain adaptation. Our approach includes a hardware-aware back-propagation orchestration agent (HAMT) and a dedicated domain-shift detector that enables active control over when and how the model is adapted (LT). Thanks to these advancements, our approach is capable of performing semantic segmentation while simultaneously adapting at more than 29FPS on a single consumer-grade GPU. Our framework's encouraging accuracy and speed trade-off is demonstrated on OnDA and SHIFT benchmarks through experimental results.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### GaitMorph: Transforming Gait by Optimally Transporting Discrete Codes
- **Authors:** Adrian Cosma, Emilian Radoi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.14713
- **Pdf link:** https://arxiv.org/pdf/2307.14713
- **Abstract**
Gait, the manner of walking, has been proven to be a reliable biometric with uses in surveillance, marketing and security. A promising new direction for the field is training gait recognition systems without explicit human annotations, through self-supervised learning approaches. Such methods are heavily reliant on strong augmentations for the same walking sequence to induce more data variability and to simulate additional walking variations. Current data augmentation schemes are heuristic and cannot provide the necessary data variation as they are only able to provide simple temporal and spatial distortions. In this work, we propose GaitMorph, a novel method to modify the walking variation for an input gait sequence. Our method entails the training of a high-compression model for gait skeleton sequences that leverages unlabelled data to construct a discrete and interpretable latent space, which preserves identity-related features. Furthermore, we propose a method based on optimal transport theory to learn latent transport maps on the discrete codebook that morph gait sequences between variations. We perform extensive experiments and show that our method is suitable to synthesize additional views for an input sequence.
### Self-Supervised Graph Transformer for Deepfake Detection
- **Authors:** Aminollah Khormali, Jiann-Shiun Yuan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2307.15019
- **Pdf link:** https://arxiv.org/pdf/2307.15019
- **Abstract**
Deepfake detection methods have shown promising results in recognizing forgeries within a given dataset, where training and testing take place on the in-distribution dataset. However, their performance deteriorates significantly when presented with unseen samples. As a result, a reliable deepfake detection system must remain impartial to forgery types, appearance, and quality for guaranteed generalizable detection performance. Despite various attempts to enhance cross-dataset generalization, the problem remains challenging, particularly when testing against common post-processing perturbations, such as video compression or blur. Hence, this study introduces a deepfake detection framework, leveraging a self-supervised pre-training model that delivers exceptional generalization ability, withstanding common corruptions and enabling feature explainability. The framework comprises three key components: a feature extractor based on vision Transformer architecture that is pre-trained via self-supervised contrastive learning methodology, a graph convolution network coupled with a Transformer discriminator, and a graph Transformer relevancy map that provides a better understanding of manipulated regions and further explains the model's decision. To assess the effectiveness of the proposed framework, several challenging experiments are conducted, including in-data distribution performance, cross-dataset, cross-manipulation generalization, and robustness against common post-production perturbations. The results achieved demonstrate the remarkable effectiveness of the proposed deepfake detection framework, surpassing the current state-of-the-art approaches.
## Keyword: RAW
### Learned Gridification for Efficient Point Cloud Processing
- **Authors:** Putri A. van der Linden, David W. Romero, Erik J. Bekkers
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2307.14354
- **Pdf link:** https://arxiv.org/pdf/2307.14354
- **Abstract**
Neural operations that rely on neighborhood information are much more expensive when deployed on point clouds than on grid data due to the irregular distances between points in a point cloud. In a grid, on the other hand, we can compute the kernel only once and reuse it for all query positions. As a result, operations that rely on neighborhood information scale much worse for point clouds than for grid data, specially for large inputs and large neighborhoods. In this work, we address the scalability issue of point cloud methods by tackling its root cause: the irregularity of the data. We propose learnable gridification as the first step in a point cloud processing pipeline to transform the point cloud into a compact, regular grid. Thanks to gridification, subsequent layers can use operations defined on regular grids, e.g., Conv3D, which scale much better than native point cloud methods. We then extend gridification to point cloud to point cloud tasks, e.g., segmentation, by adding a learnable de-gridification step at the end of the point cloud processing pipeline to map the compact, regular grid back to its original point cloud form. Through theoretical and empirical analysis, we show that gridified networks scale better in terms of memory and time than networks directly applied on raw point cloud data, while being able to achieve competitive results. Our code is publicly available at https://github.com/computri/gridifier.
### FakeTracer: Proactively Defending Against Face-swap DeepFakes via Implanting Traces in Training
- **Authors:** Pu Sun, Honggang Qi, Yuezun Li, Siwei Lyu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.14593
- **Pdf link:** https://arxiv.org/pdf/2307.14593
- **Abstract**
Face-swap DeepFake is an emerging AI-based face forgery technique that can replace the original face in a video with a generated face of the target identity while retaining consistent facial attributes such as expression and orientation. Due to the high privacy of faces, the misuse of this technique can raise severe social concerns, drawing tremendous attention to defend against DeepFakes recently. In this paper, we describe a new proactive defense method called FakeTracer to expose face-swap DeepFakes via implanting traces in training. Compared to general face-synthesis DeepFake, the face-swap DeepFake is more complex as it involves identity change, is subjected to the encoding-decoding process, and is trained unsupervised, increasing the difficulty of implanting traces into the training phase. To effectively defend against face-swap DeepFake, we design two types of traces, sustainable trace (STrace) and erasable trace (ETrace), to be added to training faces. During the training, these manipulated faces affect the learning of the face-swap DeepFake model, enabling it to generate faces that only contain sustainable traces. In light of these two traces, our method can effectively expose DeepFakes by identifying them. Extensive experiments are conducted on the Celeb-DF dataset, compared with recent passive and proactive defense methods, and are studied thoroughly regarding various factors, corroborating the efficacy of our method on defending against face-swap DeepFake.
### Exploring Annotation-free Image Captioning with Retrieval-augmented Pseudo Sentence Generation
- **Authors:** Zhiyuan Li, Dongnan Liu, Heng Wang, Chaoyi Zhang, Weidong Cai
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.14750
- **Pdf link:** https://arxiv.org/pdf/2307.14750
- **Abstract**
Training an image captioner without annotated image-sentence pairs has gained traction in recent years. Previous approaches can be categorized into two strategies: crawling sentences from mismatching corpora and aligning them with the given images as pseudo annotations, or pre-training the captioner using external image-text pairs. However, the aligning setting seems to reach its performance limit due to the quality problem of pairs, and pre-training requires significant computational resources. To address these challenges, we propose a new strategy ``LPM + retrieval-augmented learning" where the prior knowledge from large pre-trained models (LPMs) is leveraged as supervision, and a retrieval process is integrated to further reinforce its effectiveness. Specifically, we introduce Retrieval-augmented Pseudo Sentence Generation (RaPSG), which adopts an efficient approach to retrieve highly relevant short region descriptions from the mismatching corpora and use them to generate a variety of pseudo sentences with distinct representations as well as high quality via LPMs. In addition, a fluency filter and a CLIP-guided training objective are further introduced to facilitate model optimization. Experimental results demonstrate that our method surpasses the SOTA pre-training model (Flamingo3B) by achieving a CIDEr score of 78.1 (+5.1) while utilizing only 0.3% of its trainable parameters (1.3B VS 33M). Importantly, our approach eliminates the need of computationally expensive pre-training processes on external datasets (e.g., the requirement of 312M image-text pairs for Flamingo3B). We further show that with a simple extension, the generated pseudo sentences can be deployed as weak supervision to boost the 1% semi-supervised image caption benchmark up to 93.4 CIDEr score (+8.9) which showcases the versatility and effectiveness of our approach.
### TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis
- **Authors:** Zihan Zhang, Richard Liu, Kfir Aberman, Rana Hanocka
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
- **Arxiv link:** https://arxiv.org/abs/2307.15042
- **Pdf link:** https://arxiv.org/pdf/2307.15042
- **Abstract**
The gradual nature of a diffusion process that synthesizes samples in small increments constitutes a key ingredient of Denoising Diffusion Probabilistic Models (DDPM), which have presented unprecedented quality in image synthesis and been recently explored in the motion domain. In this work, we propose to adapt the gradual diffusion concept (operating along a diffusion time-axis) into the temporal-axis of the motion sequence. Our key idea is to extend the DDPM framework to support temporally varying denoising, thereby entangling the two axes. Using our special formulation, we iteratively denoise a motion buffer that contains a set of increasingly-noised poses, which auto-regressively produces an arbitrarily long stream of frames. With a stationary diffusion time-axis, in each diffusion step we increment only the temporal-axis of the motion such that the framework produces a new, clean frame which is removed from the beginning of the buffer, followed by a newly drawn noise vector that is appended to it. This new mechanism paves the way towards a new framework for long-term motion synthesis with applications to character animation and other domains.
### The RoboDepth Challenge: Methods and Advancements Towards Robust Depth Estimation
- **Authors:** Lingdong Kong, Yaru Niu, Shaoyuan Xie, Hanjiang Hu, Lai Xing Ng, Benoit R. Cottereau, Ding Zhao, Liangjun Zhang, Hesheng Wang, Wei Tsang Ooi, Ruijie Zhu, Ziyang Song, Li Liu, Tianzhu Zhang, Jun Yu, Mohan Jing, Pengwei Li, Xiaohua Qi, Cheng Jin, Yingfeng Chen, Jie Hou, Jie Zhang, Zhen Kan, Qiang Ling, Liang Peng, Minglei Li, Di Xu, Changpeng Yang, Yuanqi Yao, Gang Wu, Jian Kuai, Xianming Liu, Junjun Jiang, Jiamian Huang, Baojun Li, Jiale Chen, Shuang Zhang, Sun Ao, Zhenyu Li, Runze Chen, Haiyong Luo, Fang Zhao, Jingze Yu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2307.15061
- **Pdf link:** https://arxiv.org/pdf/2307.15061
- **Abstract**
Accurate depth estimation under out-of-distribution (OoD) scenarios, such as adverse weather conditions, sensor failure, and noise contamination, is desirable for safety-critical applications. Existing depth estimation systems, however, suffer inevitably from real-world corruptions and perturbations and are struggled to provide reliable depth predictions under such cases. In this paper, we summarize the winning solutions from the RoboDepth Challenge -- an academic competition designed to facilitate and advance robust OoD depth estimation. This challenge was developed based on the newly established KITTI-C and NYUDepth2-C benchmarks. We hosted two stand-alone tracks, with an emphasis on robust self-supervised and robust fully-supervised depth estimation, respectively. Out of more than two hundred participants, nine unique and top-performing solutions have appeared, with novel designs ranging from the following aspects: spatial- and frequency-domain augmentations, masked image modeling, image restoration and super-resolution, adversarial training, diffusion-based noise suppression, vision-language pre-training, learned model ensembling, and hierarchical feature enhancement. Extensive experimental analyses along with insightful observations are drawn to better understand the rationale behind each design. We hope this challenge could lay a solid foundation for future research on robust and reliable depth estimation and beyond. The datasets, competition toolkit, workshop recordings, and source code from the winning teams are publicly available on the challenge website.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Fri, 28 Jul 23 - ## Keyword: events
### A Memory-Augmented Multi-Task Collaborative Framework for Unsupervised Traffic Accident Detection in Driving Videos
- **Authors:** Rongqin Liang, Yuanman Li, Yingxin Yi, Jiantao Zhou, Xia Li
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.14575
- **Pdf link:** https://arxiv.org/pdf/2307.14575
- **Abstract**
Identifying traffic accidents in driving videos is crucial to ensuring the safety of autonomous driving and driver assistance systems. To address the potential danger caused by the long-tailed distribution of driving events, existing traffic accident detection (TAD) methods mainly rely on unsupervised learning. However, TAD is still challenging due to the rapid movement of cameras and dynamic scenes in driving scenarios. Existing unsupervised TAD methods mainly rely on a single pretext task, i.e., an appearance-based or future object localization task, to detect accidents. However, appearance-based approaches are easily disturbed by the rapid movement of the camera and changes in illumination, which significantly reduce the performance of traffic accident detection. Methods based on future object localization may fail to capture appearance changes in video frames, making it difficult to detect ego-involved accidents (e.g., out of control of the ego-vehicle). In this paper, we propose a novel memory-augmented multi-task collaborative framework (MAMTCF) for unsupervised traffic accident detection in driving videos. Different from previous approaches, our method can more accurately detect both ego-involved and non-ego accidents by simultaneously modeling appearance changes and object motions in video frames through the collaboration of optical flow reconstruction and future object localization tasks. Further, we introduce a memory-augmented motion representation mechanism to fully explore the interrelation between different types of motion representations and exploit the high-level features of normal traffic patterns stored in memory to augment motion representations, thus enlarging the difference from anomalies. Experimental results on recently published large-scale dataset demonstrate that our method achieves better performance compared to previous state-of-the-art approaches.
### To Adapt or Not to Adapt? Real-Time Adaptation for Semantic Segmentation
- **Authors:** Marc Botet Colomer, Pier Luigi Dovesi, Theodoros Panagiotakopoulos, Joao Frederico Carvalho, Linus Härenstam-Nielsen, Hossein Azizpour, Hedvig Kjellström, Daniel Cremers, Matteo Poggi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.15063
- **Pdf link:** https://arxiv.org/pdf/2307.15063
- **Abstract**
The goal of Online Domain Adaptation for semantic segmentation is to handle unforeseeable domain changes that occur during deployment, like sudden weather events. However, the high computational costs associated with brute-force adaptation make this paradigm unfeasible for real-world applications. In this paper we propose HAMLET, a Hardware-Aware Modular Least Expensive Training framework for real-time domain adaptation. Our approach includes a hardware-aware back-propagation orchestration agent (HAMT) and a dedicated domain-shift detector that enables active control over when and how the model is adapted (LT). Thanks to these advancements, our approach is capable of performing semantic segmentation while simultaneously adapting at more than 29FPS on a single consumer-grade GPU. Our framework's encouraging accuracy and speed trade-off is demonstrated on OnDA and SHIFT benchmarks through experimental results.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### GaitMorph: Transforming Gait by Optimally Transporting Discrete Codes
- **Authors:** Adrian Cosma, Emilian Radoi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.14713
- **Pdf link:** https://arxiv.org/pdf/2307.14713
- **Abstract**
Gait, the manner of walking, has been proven to be a reliable biometric with uses in surveillance, marketing and security. A promising new direction for the field is training gait recognition systems without explicit human annotations, through self-supervised learning approaches. Such methods are heavily reliant on strong augmentations for the same walking sequence to induce more data variability and to simulate additional walking variations. Current data augmentation schemes are heuristic and cannot provide the necessary data variation as they are only able to provide simple temporal and spatial distortions. In this work, we propose GaitMorph, a novel method to modify the walking variation for an input gait sequence. Our method entails the training of a high-compression model for gait skeleton sequences that leverages unlabelled data to construct a discrete and interpretable latent space, which preserves identity-related features. Furthermore, we propose a method based on optimal transport theory to learn latent transport maps on the discrete codebook that morph gait sequences between variations. We perform extensive experiments and show that our method is suitable to synthesize additional views for an input sequence.
### Self-Supervised Graph Transformer for Deepfake Detection
- **Authors:** Aminollah Khormali, Jiann-Shiun Yuan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2307.15019
- **Pdf link:** https://arxiv.org/pdf/2307.15019
- **Abstract**
Deepfake detection methods have shown promising results in recognizing forgeries within a given dataset, where training and testing take place on the in-distribution dataset. However, their performance deteriorates significantly when presented with unseen samples. As a result, a reliable deepfake detection system must remain impartial to forgery types, appearance, and quality for guaranteed generalizable detection performance. Despite various attempts to enhance cross-dataset generalization, the problem remains challenging, particularly when testing against common post-processing perturbations, such as video compression or blur. Hence, this study introduces a deepfake detection framework, leveraging a self-supervised pre-training model that delivers exceptional generalization ability, withstanding common corruptions and enabling feature explainability. The framework comprises three key components: a feature extractor based on vision Transformer architecture that is pre-trained via self-supervised contrastive learning methodology, a graph convolution network coupled with a Transformer discriminator, and a graph Transformer relevancy map that provides a better understanding of manipulated regions and further explains the model's decision. To assess the effectiveness of the proposed framework, several challenging experiments are conducted, including in-data distribution performance, cross-dataset, cross-manipulation generalization, and robustness against common post-production perturbations. The results achieved demonstrate the remarkable effectiveness of the proposed deepfake detection framework, surpassing the current state-of-the-art approaches.
## Keyword: RAW
### Learned Gridification for Efficient Point Cloud Processing
- **Authors:** Putri A. van der Linden, David W. Romero, Erik J. Bekkers
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2307.14354
- **Pdf link:** https://arxiv.org/pdf/2307.14354
- **Abstract**
Neural operations that rely on neighborhood information are much more expensive when deployed on point clouds than on grid data due to the irregular distances between points in a point cloud. In a grid, on the other hand, we can compute the kernel only once and reuse it for all query positions. As a result, operations that rely on neighborhood information scale much worse for point clouds than for grid data, specially for large inputs and large neighborhoods. In this work, we address the scalability issue of point cloud methods by tackling its root cause: the irregularity of the data. We propose learnable gridification as the first step in a point cloud processing pipeline to transform the point cloud into a compact, regular grid. Thanks to gridification, subsequent layers can use operations defined on regular grids, e.g., Conv3D, which scale much better than native point cloud methods. We then extend gridification to point cloud to point cloud tasks, e.g., segmentation, by adding a learnable de-gridification step at the end of the point cloud processing pipeline to map the compact, regular grid back to its original point cloud form. Through theoretical and empirical analysis, we show that gridified networks scale better in terms of memory and time than networks directly applied on raw point cloud data, while being able to achieve competitive results. Our code is publicly available at https://github.com/computri/gridifier.
### FakeTracer: Proactively Defending Against Face-swap DeepFakes via Implanting Traces in Training
- **Authors:** Pu Sun, Honggang Qi, Yuezun Li, Siwei Lyu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.14593
- **Pdf link:** https://arxiv.org/pdf/2307.14593
- **Abstract**
Face-swap DeepFake is an emerging AI-based face forgery technique that can replace the original face in a video with a generated face of the target identity while retaining consistent facial attributes such as expression and orientation. Due to the high privacy of faces, the misuse of this technique can raise severe social concerns, drawing tremendous attention to defend against DeepFakes recently. In this paper, we describe a new proactive defense method called FakeTracer to expose face-swap DeepFakes via implanting traces in training. Compared to general face-synthesis DeepFake, the face-swap DeepFake is more complex as it involves identity change, is subjected to the encoding-decoding process, and is trained unsupervised, increasing the difficulty of implanting traces into the training phase. To effectively defend against face-swap DeepFake, we design two types of traces, sustainable trace (STrace) and erasable trace (ETrace), to be added to training faces. During the training, these manipulated faces affect the learning of the face-swap DeepFake model, enabling it to generate faces that only contain sustainable traces. In light of these two traces, our method can effectively expose DeepFakes by identifying them. Extensive experiments are conducted on the Celeb-DF dataset, compared with recent passive and proactive defense methods, and are studied thoroughly regarding various factors, corroborating the efficacy of our method on defending against face-swap DeepFake.
### Exploring Annotation-free Image Captioning with Retrieval-augmented Pseudo Sentence Generation
- **Authors:** Zhiyuan Li, Dongnan Liu, Heng Wang, Chaoyi Zhang, Weidong Cai
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.14750
- **Pdf link:** https://arxiv.org/pdf/2307.14750
- **Abstract**
Training an image captioner without annotated image-sentence pairs has gained traction in recent years. Previous approaches can be categorized into two strategies: crawling sentences from mismatching corpora and aligning them with the given images as pseudo annotations, or pre-training the captioner using external image-text pairs. However, the aligning setting seems to reach its performance limit due to the quality problem of pairs, and pre-training requires significant computational resources. To address these challenges, we propose a new strategy ``LPM + retrieval-augmented learning" where the prior knowledge from large pre-trained models (LPMs) is leveraged as supervision, and a retrieval process is integrated to further reinforce its effectiveness. Specifically, we introduce Retrieval-augmented Pseudo Sentence Generation (RaPSG), which adopts an efficient approach to retrieve highly relevant short region descriptions from the mismatching corpora and use them to generate a variety of pseudo sentences with distinct representations as well as high quality via LPMs. In addition, a fluency filter and a CLIP-guided training objective are further introduced to facilitate model optimization. Experimental results demonstrate that our method surpasses the SOTA pre-training model (Flamingo3B) by achieving a CIDEr score of 78.1 (+5.1) while utilizing only 0.3% of its trainable parameters (1.3B VS 33M). Importantly, our approach eliminates the need of computationally expensive pre-training processes on external datasets (e.g., the requirement of 312M image-text pairs for Flamingo3B). We further show that with a simple extension, the generated pseudo sentences can be deployed as weak supervision to boost the 1% semi-supervised image caption benchmark up to 93.4 CIDEr score (+8.9) which showcases the versatility and effectiveness of our approach.
### TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis
- **Authors:** Zihan Zhang, Richard Liu, Kfir Aberman, Rana Hanocka
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
- **Arxiv link:** https://arxiv.org/abs/2307.15042
- **Pdf link:** https://arxiv.org/pdf/2307.15042
- **Abstract**
The gradual nature of a diffusion process that synthesizes samples in small increments constitutes a key ingredient of Denoising Diffusion Probabilistic Models (DDPM), which have presented unprecedented quality in image synthesis and been recently explored in the motion domain. In this work, we propose to adapt the gradual diffusion concept (operating along a diffusion time-axis) into the temporal-axis of the motion sequence. Our key idea is to extend the DDPM framework to support temporally varying denoising, thereby entangling the two axes. Using our special formulation, we iteratively denoise a motion buffer that contains a set of increasingly-noised poses, which auto-regressively produces an arbitrarily long stream of frames. With a stationary diffusion time-axis, in each diffusion step we increment only the temporal-axis of the motion such that the framework produces a new, clean frame which is removed from the beginning of the buffer, followed by a newly drawn noise vector that is appended to it. This new mechanism paves the way towards a new framework for long-term motion synthesis with applications to character animation and other domains.
### The RoboDepth Challenge: Methods and Advancements Towards Robust Depth Estimation
- **Authors:** Lingdong Kong, Yaru Niu, Shaoyuan Xie, Hanjiang Hu, Lai Xing Ng, Benoit R. Cottereau, Ding Zhao, Liangjun Zhang, Hesheng Wang, Wei Tsang Ooi, Ruijie Zhu, Ziyang Song, Li Liu, Tianzhu Zhang, Jun Yu, Mohan Jing, Pengwei Li, Xiaohua Qi, Cheng Jin, Yingfeng Chen, Jie Hou, Jie Zhang, Zhen Kan, Qiang Ling, Liang Peng, Minglei Li, Di Xu, Changpeng Yang, Yuanqi Yao, Gang Wu, Jian Kuai, Xianming Liu, Junjun Jiang, Jiamian Huang, Baojun Li, Jiale Chen, Shuang Zhang, Sun Ao, Zhenyu Li, Runze Chen, Haiyong Luo, Fang Zhao, Jingze Yu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2307.15061
- **Pdf link:** https://arxiv.org/pdf/2307.15061
- **Abstract**
Accurate depth estimation under out-of-distribution (OoD) scenarios, such as adverse weather conditions, sensor failure, and noise contamination, is desirable for safety-critical applications. Existing depth estimation systems, however, suffer inevitably from real-world corruptions and perturbations and are struggled to provide reliable depth predictions under such cases. In this paper, we summarize the winning solutions from the RoboDepth Challenge -- an academic competition designed to facilitate and advance robust OoD depth estimation. This challenge was developed based on the newly established KITTI-C and NYUDepth2-C benchmarks. We hosted two stand-alone tracks, with an emphasis on robust self-supervised and robust fully-supervised depth estimation, respectively. Out of more than two hundred participants, nine unique and top-performing solutions have appeared, with novel designs ranging from the following aspects: spatial- and frequency-domain augmentations, masked image modeling, image restoration and super-resolution, adversarial training, diffusion-based noise suppression, vision-language pre-training, learned model ensembling, and hierarchical feature enhancement. Extensive experimental analyses along with insightful observations are drawn to better understand the rationale behind each design. We hope this challenge could lay a solid foundation for future research on robust and reliable depth estimation and beyond. The datasets, competition toolkit, workshop recordings, and source code from the winning teams are publicly available on the challenge website.
## Keyword: raw image
There is no result
|
process
|
new submissions for fri jul keyword events a memory augmented multi task collaborative framework for unsupervised traffic accident detection in driving videos authors rongqin liang yuanman li yingxin yi jiantao zhou xia li subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract identifying traffic accidents in driving videos is crucial to ensuring the safety of autonomous driving and driver assistance systems to address the potential danger caused by the long tailed distribution of driving events existing traffic accident detection tad methods mainly rely on unsupervised learning however tad is still challenging due to the rapid movement of cameras and dynamic scenes in driving scenarios existing unsupervised tad methods mainly rely on a single pretext task i e an appearance based or future object localization task to detect accidents however appearance based approaches are easily disturbed by the rapid movement of the camera and changes in illumination which significantly reduce the performance of traffic accident detection methods based on future object localization may fail to capture appearance changes in video frames making it difficult to detect ego involved accidents e g out of control of the ego vehicle in this paper we propose a novel memory augmented multi task collaborative framework mamtcf for unsupervised traffic accident detection in driving videos different from previous approaches our method can more accurately detect both ego involved and non ego accidents by simultaneously modeling appearance changes and object motions in video frames through the collaboration of optical flow reconstruction and future object localization tasks further we introduce a memory augmented motion representation mechanism to fully explore the interrelation between different types of motion representations and exploit the high level features of normal traffic patterns stored in memory to augment motion representations thus enlarging the difference from anomalies experimental results on recently published large scale dataset demonstrate that our method achieves better performance compared to previous state of the art approaches to adapt or not to adapt real time adaptation for semantic segmentation authors marc botet colomer pier luigi dovesi theodoros panagiotakopoulos joao frederico carvalho linus härenstam nielsen hossein azizpour hedvig kjellström daniel cremers matteo poggi subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the goal of online domain adaptation for semantic segmentation is to handle unforeseeable domain changes that occur during deployment like sudden weather events however the high computational costs associated with brute force adaptation make this paradigm unfeasible for real world applications in this paper we propose hamlet a hardware aware modular least expensive training framework for real time domain adaptation our approach includes a hardware aware back propagation orchestration agent hamt and a dedicated domain shift detector that enables active control over when and how the model is adapted lt thanks to these advancements our approach is capable of performing semantic segmentation while simultaneously adapting at more than on a single consumer grade gpu our framework s encouraging accuracy and speed trade off is demonstrated on onda and shift benchmarks through experimental results keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression gaitmorph transforming gait by optimally transporting discrete codes authors adrian cosma emilian radoi subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract gait the manner of walking has been proven to be a reliable biometric with uses in surveillance marketing and security a promising new direction for the field is training gait recognition systems without explicit human annotations through self supervised learning approaches such methods are heavily reliant on strong augmentations for the same walking sequence to induce more data variability and to simulate additional walking variations current data augmentation schemes are heuristic and cannot provide the necessary data variation as they are only able to provide simple temporal and spatial distortions in this work we propose gaitmorph a novel method to modify the walking variation for an input gait sequence our method entails the training of a high compression model for gait skeleton sequences that leverages unlabelled data to construct a discrete and interpretable latent space which preserves identity related features furthermore we propose a method based on optimal transport theory to learn latent transport maps on the discrete codebook that morph gait sequences between variations we perform extensive experiments and show that our method is suitable to synthesize additional views for an input sequence self supervised graph transformer for deepfake detection authors aminollah khormali jiann shiun yuan subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract deepfake detection methods have shown promising results in recognizing forgeries within a given dataset where training and testing take place on the in distribution dataset however their performance deteriorates significantly when presented with unseen samples as a result a reliable deepfake detection system must remain impartial to forgery types appearance and quality for guaranteed generalizable detection performance despite various attempts to enhance cross dataset generalization the problem remains challenging particularly when testing against common post processing perturbations such as video compression or blur hence this study introduces a deepfake detection framework leveraging a self supervised pre training model that delivers exceptional generalization ability withstanding common corruptions and enabling feature explainability the framework comprises three key components a feature extractor based on vision transformer architecture that is pre trained via self supervised contrastive learning methodology a graph convolution network coupled with a transformer discriminator and a graph transformer relevancy map that provides a better understanding of manipulated regions and further explains the model s decision to assess the effectiveness of the proposed framework several challenging experiments are conducted including in data distribution performance cross dataset cross manipulation generalization and robustness against common post production perturbations the results achieved demonstrate the remarkable effectiveness of the proposed deepfake detection framework surpassing the current state of the art approaches keyword raw learned gridification for efficient point cloud processing authors putri a van der linden david w romero erik j bekkers subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract neural operations that rely on neighborhood information are much more expensive when deployed on point clouds than on grid data due to the irregular distances between points in a point cloud in a grid on the other hand we can compute the kernel only once and reuse it for all query positions as a result operations that rely on neighborhood information scale much worse for point clouds than for grid data specially for large inputs and large neighborhoods in this work we address the scalability issue of point cloud methods by tackling its root cause the irregularity of the data we propose learnable gridification as the first step in a point cloud processing pipeline to transform the point cloud into a compact regular grid thanks to gridification subsequent layers can use operations defined on regular grids e g which scale much better than native point cloud methods we then extend gridification to point cloud to point cloud tasks e g segmentation by adding a learnable de gridification step at the end of the point cloud processing pipeline to map the compact regular grid back to its original point cloud form through theoretical and empirical analysis we show that gridified networks scale better in terms of memory and time than networks directly applied on raw point cloud data while being able to achieve competitive results our code is publicly available at faketracer proactively defending against face swap deepfakes via implanting traces in training authors pu sun honggang qi yuezun li siwei lyu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract face swap deepfake is an emerging ai based face forgery technique that can replace the original face in a video with a generated face of the target identity while retaining consistent facial attributes such as expression and orientation due to the high privacy of faces the misuse of this technique can raise severe social concerns drawing tremendous attention to defend against deepfakes recently in this paper we describe a new proactive defense method called faketracer to expose face swap deepfakes via implanting traces in training compared to general face synthesis deepfake the face swap deepfake is more complex as it involves identity change is subjected to the encoding decoding process and is trained unsupervised increasing the difficulty of implanting traces into the training phase to effectively defend against face swap deepfake we design two types of traces sustainable trace strace and erasable trace etrace to be added to training faces during the training these manipulated faces affect the learning of the face swap deepfake model enabling it to generate faces that only contain sustainable traces in light of these two traces our method can effectively expose deepfakes by identifying them extensive experiments are conducted on the celeb df dataset compared with recent passive and proactive defense methods and are studied thoroughly regarding various factors corroborating the efficacy of our method on defending against face swap deepfake exploring annotation free image captioning with retrieval augmented pseudo sentence generation authors zhiyuan li dongnan liu heng wang chaoyi zhang weidong cai subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract training an image captioner without annotated image sentence pairs has gained traction in recent years previous approaches can be categorized into two strategies crawling sentences from mismatching corpora and aligning them with the given images as pseudo annotations or pre training the captioner using external image text pairs however the aligning setting seems to reach its performance limit due to the quality problem of pairs and pre training requires significant computational resources to address these challenges we propose a new strategy lpm retrieval augmented learning where the prior knowledge from large pre trained models lpms is leveraged as supervision and a retrieval process is integrated to further reinforce its effectiveness specifically we introduce retrieval augmented pseudo sentence generation rapsg which adopts an efficient approach to retrieve highly relevant short region descriptions from the mismatching corpora and use them to generate a variety of pseudo sentences with distinct representations as well as high quality via lpms in addition a fluency filter and a clip guided training objective are further introduced to facilitate model optimization experimental results demonstrate that our method surpasses the sota pre training model by achieving a cider score of while utilizing only of its trainable parameters vs importantly our approach eliminates the need of computationally expensive pre training processes on external datasets e g the requirement of image text pairs for we further show that with a simple extension the generated pseudo sentences can be deployed as weak supervision to boost the semi supervised image caption benchmark up to cider score which showcases the versatility and effectiveness of our approach tedi temporally entangled diffusion for long term motion synthesis authors zihan zhang richard liu kfir aberman rana hanocka subjects computer vision and pattern recognition cs cv graphics cs gr arxiv link pdf link abstract the gradual nature of a diffusion process that synthesizes samples in small increments constitutes a key ingredient of denoising diffusion probabilistic models ddpm which have presented unprecedented quality in image synthesis and been recently explored in the motion domain in this work we propose to adapt the gradual diffusion concept operating along a diffusion time axis into the temporal axis of the motion sequence our key idea is to extend the ddpm framework to support temporally varying denoising thereby entangling the two axes using our special formulation we iteratively denoise a motion buffer that contains a set of increasingly noised poses which auto regressively produces an arbitrarily long stream of frames with a stationary diffusion time axis in each diffusion step we increment only the temporal axis of the motion such that the framework produces a new clean frame which is removed from the beginning of the buffer followed by a newly drawn noise vector that is appended to it this new mechanism paves the way towards a new framework for long term motion synthesis with applications to character animation and other domains the robodepth challenge methods and advancements towards robust depth estimation authors lingdong kong yaru niu shaoyuan xie hanjiang hu lai xing ng benoit r cottereau ding zhao liangjun zhang hesheng wang wei tsang ooi ruijie zhu ziyang song li liu tianzhu zhang jun yu mohan jing pengwei li xiaohua qi cheng jin yingfeng chen jie hou jie zhang zhen kan qiang ling liang peng minglei li di xu changpeng yang yuanqi yao gang wu jian kuai xianming liu junjun jiang jiamian huang baojun li jiale chen shuang zhang sun ao zhenyu li runze chen haiyong luo fang zhao jingze yu subjects computer vision and pattern recognition cs cv robotics cs ro arxiv link pdf link abstract accurate depth estimation under out of distribution ood scenarios such as adverse weather conditions sensor failure and noise contamination is desirable for safety critical applications existing depth estimation systems however suffer inevitably from real world corruptions and perturbations and are struggled to provide reliable depth predictions under such cases in this paper we summarize the winning solutions from the robodepth challenge an academic competition designed to facilitate and advance robust ood depth estimation this challenge was developed based on the newly established kitti c and c benchmarks we hosted two stand alone tracks with an emphasis on robust self supervised and robust fully supervised depth estimation respectively out of more than two hundred participants nine unique and top performing solutions have appeared with novel designs ranging from the following aspects spatial and frequency domain augmentations masked image modeling image restoration and super resolution adversarial training diffusion based noise suppression vision language pre training learned model ensembling and hierarchical feature enhancement extensive experimental analyses along with insightful observations are drawn to better understand the rationale behind each design we hope this challenge could lay a solid foundation for future research on robust and reliable depth estimation and beyond the datasets competition toolkit workshop recordings and source code from the winning teams are publicly available on the challenge website keyword raw image there is no result
| 1
|
9,196
| 12,231,474,320
|
IssuesEvent
|
2020-05-04 07:51:10
|
naoki-shigehisa/paper
|
https://api.github.com/repos/naoki-shigehisa/paper
|
closed
|
Decomposing feature-level variation with Covariate Gaussian Process Latent Variable Models
|
2018 Gaussian Process
|
## 0. 論文
タイトル:[Decomposing feature-level variation with Covariate Gaussian Process Latent Variable Models](https://arxiv.org/abs/1810.06983)
著者:

arXiv投稿日:2018/10/16
学会/ジャーナル:ICML 2019
## 1. どんなもの?
共変量や変数間の非線形相互作用を考慮して次元削減を行う、c-GPLVMの提案
## 2. 先行研究と比べてどこがすごい?
共変量の変化による変動もキャプチャできる(下図の上:GPLVM、下:c-GPLVM)

x:共変量 z:潜在変数
## 3. 技術や手法のキモはどこ?
z(潜在変数)とx(共変量)の空間で定義されるマッピングを学習する
(zの関数としてxとyの両方をモデル化するのではない)
共同空間でARDカーネルを定義する

ADD(付加)モデルまたはINT(相互作用)モデルのどちらか(どちらも)を利用

## 4. どうやって有効だと検証した?
2次元のtoy dataで検証

Cancer Genome Atlas cohort(乳癌のデータセット)で検証

## 5. 議論はある?
難しい
使い所も難しい
## 6. 次に読むべき論文は?
|
1.0
|
Decomposing feature-level variation with Covariate Gaussian Process Latent Variable Models - ## 0. 論文
タイトル:[Decomposing feature-level variation with Covariate Gaussian Process Latent Variable Models](https://arxiv.org/abs/1810.06983)
著者:

arXiv投稿日:2018/10/16
学会/ジャーナル:ICML 2019
## 1. どんなもの?
共変量や変数間の非線形相互作用を考慮して次元削減を行う、c-GPLVMの提案
## 2. 先行研究と比べてどこがすごい?
共変量の変化による変動もキャプチャできる(下図の上:GPLVM、下:c-GPLVM)

x:共変量 z:潜在変数
## 3. 技術や手法のキモはどこ?
z(潜在変数)とx(共変量)の空間で定義されるマッピングを学習する
(zの関数としてxとyの両方をモデル化するのではない)
共同空間でARDカーネルを定義する

ADD(付加)モデルまたはINT(相互作用)モデルのどちらか(どちらも)を利用

## 4. どうやって有効だと検証した?
2次元のtoy dataで検証

Cancer Genome Atlas cohort(乳癌のデータセット)で検証

## 5. 議論はある?
難しい
使い所も難しい
## 6. 次に読むべき論文は?
|
process
|
decomposing feature level variation with covariate gaussian process latent variable models 論文 タイトル: 著者: arxiv投稿日: 学会 ジャーナル:icml どんなもの? 共変量や変数間の非線形相互作用を考慮して次元削減を行う、c gplvmの提案 先行研究と比べてどこがすごい? 共変量の変化による変動もキャプチャできる 下図の上:gplvm、下:c gplvm x:共変量 z:潜在変数 技術や手法のキモはどこ? z 潜在変数 とx 共変量 の空間で定義されるマッピングを学習する zの関数としてxとyの両方をモデル化するのではない 共同空間でardカーネルを定義する add 付加 モデルまたはint 相互作用 モデルのどちらか どちらも を利用 どうやって有効だと検証した? dataで検証 cancer genome atlas cohort 乳癌のデータセット で検証 議論はある? 難しい 使い所も難しい 次に読むべき論文は?
| 1
|
71,925
| 7,269,125,453
|
IssuesEvent
|
2018-02-20 12:37:17
|
JuliaLang/julia
|
https://api.github.com/repos/JuliaLang/julia
|
closed
|
InterruptExceptions should probably just abort Pkg.test()
|
testsystem
|
It is quite difficult to abort a running Pkg.Test because the test framework happily accepts a InterruptException as simply a failed test and moves on to the next one. Holding Ctrl+C usually gets it to die but then so does the current session.
|
1.0
|
InterruptExceptions should probably just abort Pkg.test() - It is quite difficult to abort a running Pkg.Test because the test framework happily accepts a InterruptException as simply a failed test and moves on to the next one. Holding Ctrl+C usually gets it to die but then so does the current session.
|
non_process
|
interruptexceptions should probably just abort pkg test it is quite difficult to abort a running pkg test because the test framework happily accepts a interruptexception as simply a failed test and moves on to the next one holding ctrl c usually gets it to die but then so does the current session
| 0
|
21,790
| 30,298,065,384
|
IssuesEvent
|
2023-07-10 02:00:08
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Mon, 10 Jul 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Freezing of Gait Prediction From Accelerometer Data Using a Simple 1D-Convolutional Neural Network -- 8th Place Solution for Kaggle's Parkinson's Freezing of Gait Prediction Competition
- **Authors:** Jan Brederecke
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03475
- **Pdf link:** https://arxiv.org/pdf/2307.03475
- **Abstract**
Freezing of Gait (FOG) is a common motor symptom in patients with Parkinson's disease (PD). During episodes of FOG, patients suddenly lose their ability to stride as intended. Patient-worn accelerometers can capture information on the patient's movement during these episodes and machine learning algorithms can potentially classify this data. The combination therefore holds the potential to detect FOG in real-time. In this work I present a simple 1-D convolutional neural network that was trained to detect FOG events in accelerometer data. Model performance was assessed by measuring the success of the model to discriminate normal movement from FOG episodes and resulted in a mean average precision of 0.356 on the private leaderboard on Kaggle. Ultimately, the model ranked 8th out of 1379 teams in the Parkinson's Freezing of Gait Prediction competition. The results underscore the potential of Deep Learning-based solutions in advancing the field of FOG detection, contributing to improved interventions and management strategies for PD patients.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### Tranfer Learning of Semantic Segmentation Methods for Identifying Buried Archaeological Structures on LiDAR Data
- **Authors:** Paolo Soleni, Wouter B. Verschoof-van der Vaart, Žiga Kokalj, Arianna Traviglia, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03512
- **Pdf link:** https://arxiv.org/pdf/2307.03512
- **Abstract**
When applying deep learning to remote sensing data in archaeological research, a notable obstacle is the limited availability of suitable datasets for training models. The application of transfer learning is frequently employed to mitigate this drawback. However, there is still a need to explore its effectiveness when applied across different archaeological datasets. This paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two LiDAR datasets. The experimental results indicate that transfer learning-based approaches in archaeology can lead to performance improvements, although a systematic enhancement has not yet been observed. We provide specific insights about the validity of such techniques that can serve as a baseline for future works.
## Keyword: ISP
### General-Purpose Multimodal Transformer meets Remote Sensing Semantic Segmentation
- **Authors:** Nhi Kieu, Kien Nguyen, Sridha Sridharan, Clinton Fookes
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03388
- **Pdf link:** https://arxiv.org/pdf/2307.03388
- **Abstract**
The advent of high-resolution multispectral/hyperspectral sensors, LiDAR DSM (Digital Surface Model) information and many others has provided us with an unprecedented wealth of data for Earth Observation. Multimodal AI seeks to exploit those complementary data sources, particularly for complex tasks like semantic segmentation. While specialized architectures have been developed, they are highly complicated via significant effort in model design, and require considerable re-engineering whenever a new modality emerges. Recent trends in general-purpose multimodal networks have shown great potential to achieve state-of-the-art performance across multiple multimodal tasks with one unified architecture. In this work, we investigate the performance of PerceiverIO, one in the general-purpose multimodal family, in the remote sensing semantic segmentation domain. Our experiments reveal that this ostensibly universal network struggles with object scale variation in remote sensing images and fails to detect the presence of cars from a top-down view. To address these issues, even with extreme class imbalance issues, we propose a spatial and volumetric learning component. Specifically, we design a UNet-inspired module that employs 3D convolution to encode vital local information and learn cross-modal features simultaneously, while reducing network computational burden via the cross-attention mechanism of PerceiverIO. The effectiveness of the proposed component is validated through extensive experiments comparing it with other methods such as 2D convolution, and dual local module (\ie the combination of Conv2D 1x1 and Conv2D 3x3 inspired by UNetFormer). The proposed method achieves competitive results with specialized architectures like UNetFormer and SwinUNet, showing its potential to minimize network architecture engineering with a minimal compromise on the performance.
### Unsupervised Hyperspectral and Multispectral Images Fusion Based on the Cycle Consistency
- **Authors:** Shuaikai Shi, Lijun Zhang, Yoann Altmann, Jie Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.03413
- **Pdf link:** https://arxiv.org/pdf/2307.03413
- **Abstract**
Hyperspectral images (HSI) with abundant spectral information reflected materials property usually perform low spatial resolution due to the hardware limits. Meanwhile, multispectral images (MSI), e.g., RGB images, have a high spatial resolution but deficient spectral signatures. Hyperspectral and multispectral image fusion can be cost-effective and efficient for acquiring both high spatial resolution and high spectral resolution images. Many of the conventional HSI and MSI fusion algorithms rely on known spatial degradation parameters, i.e., point spread function, spectral degradation parameters, spectral response function, or both of them. Another class of deep learning-based models relies on the ground truth of high spatial resolution HSI and needs large amounts of paired training images when working in a supervised manner. Both of these models are limited in practical fusion scenarios. In this paper, we propose an unsupervised HSI and MSI fusion model based on the cycle consistency, called CycFusion. The CycFusion learns the domain transformation between low spatial resolution HSI (LrHSI) and high spatial resolution MSI (HrMSI), and the desired high spatial resolution HSI (HrHSI) are considered to be intermediate feature maps in the transformation networks. The CycFusion can be trained with the objective functions of marginal matching in single transform and cycle consistency in double transforms. Moreover, the estimated PSF and SRF are embedded in the model as the pre-training weights, which further enhances the practicality of our proposed model. Experiments conducted on several datasets show that our proposed model outperforms all compared unsupervised fusion methods. The codes of this paper will be available at this address: https: //github.com/shuaikaishi/CycFusion for reproducibility.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### All in One: Exploring Unified Vision-Language Tracking with Multi-Modal Alignment
- **Authors:** Chunhui Zhang, Xin Sun, Li Liu, Yiqian Yang, Qiong Liu, Xi Zhou, Yanfeng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03373
- **Pdf link:** https://arxiv.org/pdf/2307.03373
- **Abstract**
Current mainstream vision-language (VL) tracking framework consists of three parts, \ie a visual feature extractor, a language feature extractor, and a fusion model. To pursue better performance, a natural modus operandi for VL tracking is employing customized and heavier unimodal encoders, and multi-modal fusion models. Albeit effective, existing VL trackers separate feature extraction and feature integration, resulting in extracted features that lack semantic guidance and have limited target-aware capability in complex scenarios, \eg similar distractors and extreme illumination. In this work, inspired by the recent success of exploring foundation models with unified architecture for both natural language and computer vision tasks, we propose an All-in-One framework, which learns joint feature extraction and interaction by adopting a unified transformer backbone. Specifically, we mix raw vision and language signals to generate language-injected vision tokens, which we then concatenate before feeding into the unified backbone architecture. This approach achieves feature integration in a unified backbone, removing the need for carefully-designed fusion modules and resulting in a more effective and efficient VL tracking framework. To further improve the learning efficiency, we introduce a multi-modal alignment module based on cross-modal and intra-modal contrastive objectives, providing more reasonable representations for the unified All-in-One transformer backbone. Extensive experiments on five benchmarks, \ie OTB99-L, TNL2K, LaSOT, LaSOT$_{\rm Ext}$ and WebUAV-3M, demonstrate the superiority of the proposed tracker against existing state-of-the-arts on VL tracking. Codes will be made publicly available.
### Tranfer Learning of Semantic Segmentation Methods for Identifying Buried Archaeological Structures on LiDAR Data
- **Authors:** Paolo Soleni, Wouter B. Verschoof-van der Vaart, Žiga Kokalj, Arianna Traviglia, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03512
- **Pdf link:** https://arxiv.org/pdf/2307.03512
- **Abstract**
When applying deep learning to remote sensing data in archaeological research, a notable obstacle is the limited availability of suitable datasets for training models. The application of transfer learning is frequently employed to mitigate this drawback. However, there is still a need to explore its effectiveness when applied across different archaeological datasets. This paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two LiDAR datasets. The experimental results indicate that transfer learning-based approaches in archaeology can lead to performance improvements, although a systematic enhancement has not yet been observed. We provide specific insights about the validity of such techniques that can serve as a baseline for future works.
### Language-free Compositional Action Generation via Decoupling Refinement
- **Authors:** Xiao Liu, Guangyi Chen, Yansong Tang, Guangrun Wang, Ser-Nam Lim
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03538
- **Pdf link:** https://arxiv.org/pdf/2307.03538
- **Abstract**
Composing simple elements into complex concepts is crucial yet challenging, especially for 3D action generation. Existing methods largely rely on extensive neural language annotations to discern composable latent semantics, a process that is often costly and labor-intensive. In this study, we introduce a novel framework to generate compositional actions without reliance on language auxiliaries. Our approach consists of three main components: Action Coupling, Conditional Action Generation, and Decoupling Refinement. Action Coupling utilizes an energy model to extract the attention masks of each sub-action, subsequently integrating two actions using these attentions to generate pseudo-training examples. Then, we employ a conditional generative model, CVAE, to learn a latent space, facilitating the diverse generation. Finally, we propose Decoupling Refinement, which leverages a self-supervised pre-trained model MAE to ensure semantic consistency between the sub-actions and compositional actions. This refinement process involves rendering generated 3D actions into 2D space, decoupling these images into two sub-segments, using the MAE model to restore the complete image from sub-segments, and constraining the recovered images to match images rendered from raw sub-actions. Due to the lack of existing datasets containing both sub-actions and compositional actions, we created two new datasets, named HumanAct-C and UESTC-C, and present a corresponding evaluation metric. Both qualitative and quantitative assessments are conducted to show our efficacy.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Mon, 10 Jul 23 - ## Keyword: events
### Freezing of Gait Prediction From Accelerometer Data Using a Simple 1D-Convolutional Neural Network -- 8th Place Solution for Kaggle's Parkinson's Freezing of Gait Prediction Competition
- **Authors:** Jan Brederecke
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03475
- **Pdf link:** https://arxiv.org/pdf/2307.03475
- **Abstract**
Freezing of Gait (FOG) is a common motor symptom in patients with Parkinson's disease (PD). During episodes of FOG, patients suddenly lose their ability to stride as intended. Patient-worn accelerometers can capture information on the patient's movement during these episodes and machine learning algorithms can potentially classify this data. The combination therefore holds the potential to detect FOG in real-time. In this work I present a simple 1-D convolutional neural network that was trained to detect FOG events in accelerometer data. Model performance was assessed by measuring the success of the model to discriminate normal movement from FOG episodes and resulted in a mean average precision of 0.356 on the private leaderboard on Kaggle. Ultimately, the model ranked 8th out of 1379 teams in the Parkinson's Freezing of Gait Prediction competition. The results underscore the potential of Deep Learning-based solutions in advancing the field of FOG detection, contributing to improved interventions and management strategies for PD patients.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### Tranfer Learning of Semantic Segmentation Methods for Identifying Buried Archaeological Structures on LiDAR Data
- **Authors:** Paolo Soleni, Wouter B. Verschoof-van der Vaart, Žiga Kokalj, Arianna Traviglia, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03512
- **Pdf link:** https://arxiv.org/pdf/2307.03512
- **Abstract**
When applying deep learning to remote sensing data in archaeological research, a notable obstacle is the limited availability of suitable datasets for training models. The application of transfer learning is frequently employed to mitigate this drawback. However, there is still a need to explore its effectiveness when applied across different archaeological datasets. This paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two LiDAR datasets. The experimental results indicate that transfer learning-based approaches in archaeology can lead to performance improvements, although a systematic enhancement has not yet been observed. We provide specific insights about the validity of such techniques that can serve as a baseline for future works.
## Keyword: ISP
### General-Purpose Multimodal Transformer meets Remote Sensing Semantic Segmentation
- **Authors:** Nhi Kieu, Kien Nguyen, Sridha Sridharan, Clinton Fookes
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03388
- **Pdf link:** https://arxiv.org/pdf/2307.03388
- **Abstract**
The advent of high-resolution multispectral/hyperspectral sensors, LiDAR DSM (Digital Surface Model) information and many others has provided us with an unprecedented wealth of data for Earth Observation. Multimodal AI seeks to exploit those complementary data sources, particularly for complex tasks like semantic segmentation. While specialized architectures have been developed, they are highly complicated via significant effort in model design, and require considerable re-engineering whenever a new modality emerges. Recent trends in general-purpose multimodal networks have shown great potential to achieve state-of-the-art performance across multiple multimodal tasks with one unified architecture. In this work, we investigate the performance of PerceiverIO, one in the general-purpose multimodal family, in the remote sensing semantic segmentation domain. Our experiments reveal that this ostensibly universal network struggles with object scale variation in remote sensing images and fails to detect the presence of cars from a top-down view. To address these issues, even with extreme class imbalance issues, we propose a spatial and volumetric learning component. Specifically, we design a UNet-inspired module that employs 3D convolution to encode vital local information and learn cross-modal features simultaneously, while reducing network computational burden via the cross-attention mechanism of PerceiverIO. The effectiveness of the proposed component is validated through extensive experiments comparing it with other methods such as 2D convolution, and dual local module (\ie the combination of Conv2D 1x1 and Conv2D 3x3 inspired by UNetFormer). The proposed method achieves competitive results with specialized architectures like UNetFormer and SwinUNet, showing its potential to minimize network architecture engineering with a minimal compromise on the performance.
### Unsupervised Hyperspectral and Multispectral Images Fusion Based on the Cycle Consistency
- **Authors:** Shuaikai Shi, Lijun Zhang, Yoann Altmann, Jie Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.03413
- **Pdf link:** https://arxiv.org/pdf/2307.03413
- **Abstract**
Hyperspectral images (HSI) with abundant spectral information reflected materials property usually perform low spatial resolution due to the hardware limits. Meanwhile, multispectral images (MSI), e.g., RGB images, have a high spatial resolution but deficient spectral signatures. Hyperspectral and multispectral image fusion can be cost-effective and efficient for acquiring both high spatial resolution and high spectral resolution images. Many of the conventional HSI and MSI fusion algorithms rely on known spatial degradation parameters, i.e., point spread function, spectral degradation parameters, spectral response function, or both of them. Another class of deep learning-based models relies on the ground truth of high spatial resolution HSI and needs large amounts of paired training images when working in a supervised manner. Both of these models are limited in practical fusion scenarios. In this paper, we propose an unsupervised HSI and MSI fusion model based on the cycle consistency, called CycFusion. The CycFusion learns the domain transformation between low spatial resolution HSI (LrHSI) and high spatial resolution MSI (HrMSI), and the desired high spatial resolution HSI (HrHSI) are considered to be intermediate feature maps in the transformation networks. The CycFusion can be trained with the objective functions of marginal matching in single transform and cycle consistency in double transforms. Moreover, the estimated PSF and SRF are embedded in the model as the pre-training weights, which further enhances the practicality of our proposed model. Experiments conducted on several datasets show that our proposed model outperforms all compared unsupervised fusion methods. The codes of this paper will be available at this address: https: //github.com/shuaikaishi/CycFusion for reproducibility.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### All in One: Exploring Unified Vision-Language Tracking with Multi-Modal Alignment
- **Authors:** Chunhui Zhang, Xin Sun, Li Liu, Yiqian Yang, Qiong Liu, Xi Zhou, Yanfeng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03373
- **Pdf link:** https://arxiv.org/pdf/2307.03373
- **Abstract**
Current mainstream vision-language (VL) tracking framework consists of three parts, \ie a visual feature extractor, a language feature extractor, and a fusion model. To pursue better performance, a natural modus operandi for VL tracking is employing customized and heavier unimodal encoders, and multi-modal fusion models. Albeit effective, existing VL trackers separate feature extraction and feature integration, resulting in extracted features that lack semantic guidance and have limited target-aware capability in complex scenarios, \eg similar distractors and extreme illumination. In this work, inspired by the recent success of exploring foundation models with unified architecture for both natural language and computer vision tasks, we propose an All-in-One framework, which learns joint feature extraction and interaction by adopting a unified transformer backbone. Specifically, we mix raw vision and language signals to generate language-injected vision tokens, which we then concatenate before feeding into the unified backbone architecture. This approach achieves feature integration in a unified backbone, removing the need for carefully-designed fusion modules and resulting in a more effective and efficient VL tracking framework. To further improve the learning efficiency, we introduce a multi-modal alignment module based on cross-modal and intra-modal contrastive objectives, providing more reasonable representations for the unified All-in-One transformer backbone. Extensive experiments on five benchmarks, \ie OTB99-L, TNL2K, LaSOT, LaSOT$_{\rm Ext}$ and WebUAV-3M, demonstrate the superiority of the proposed tracker against existing state-of-the-arts on VL tracking. Codes will be made publicly available.
### Tranfer Learning of Semantic Segmentation Methods for Identifying Buried Archaeological Structures on LiDAR Data
- **Authors:** Paolo Soleni, Wouter B. Verschoof-van der Vaart, Žiga Kokalj, Arianna Traviglia, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.03512
- **Pdf link:** https://arxiv.org/pdf/2307.03512
- **Abstract**
When applying deep learning to remote sensing data in archaeological research, a notable obstacle is the limited availability of suitable datasets for training models. The application of transfer learning is frequently employed to mitigate this drawback. However, there is still a need to explore its effectiveness when applied across different archaeological datasets. This paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two LiDAR datasets. The experimental results indicate that transfer learning-based approaches in archaeology can lead to performance improvements, although a systematic enhancement has not yet been observed. We provide specific insights about the validity of such techniques that can serve as a baseline for future works.
### Language-free Compositional Action Generation via Decoupling Refinement
- **Authors:** Xiao Liu, Guangyi Chen, Yansong Tang, Guangrun Wang, Ser-Nam Lim
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.03538
- **Pdf link:** https://arxiv.org/pdf/2307.03538
- **Abstract**
Composing simple elements into complex concepts is crucial yet challenging, especially for 3D action generation. Existing methods largely rely on extensive neural language annotations to discern composable latent semantics, a process that is often costly and labor-intensive. In this study, we introduce a novel framework to generate compositional actions without reliance on language auxiliaries. Our approach consists of three main components: Action Coupling, Conditional Action Generation, and Decoupling Refinement. Action Coupling utilizes an energy model to extract the attention masks of each sub-action, subsequently integrating two actions using these attentions to generate pseudo-training examples. Then, we employ a conditional generative model, CVAE, to learn a latent space, facilitating the diverse generation. Finally, we propose Decoupling Refinement, which leverages a self-supervised pre-trained model MAE to ensure semantic consistency between the sub-actions and compositional actions. This refinement process involves rendering generated 3D actions into 2D space, decoupling these images into two sub-segments, using the MAE model to restore the complete image from sub-segments, and constraining the recovered images to match images rendered from raw sub-actions. Due to the lack of existing datasets containing both sub-actions and compositional actions, we created two new datasets, named HumanAct-C and UESTC-C, and present a corresponding evaluation metric. Both qualitative and quantitative assessments are conducted to show our efficacy.
## Keyword: raw image
There is no result
|
process
|
new submissions for mon jul keyword events freezing of gait prediction from accelerometer data using a simple convolutional neural network place solution for kaggle s parkinson s freezing of gait prediction competition authors jan brederecke subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract freezing of gait fog is a common motor symptom in patients with parkinson s disease pd during episodes of fog patients suddenly lose their ability to stride as intended patient worn accelerometers can capture information on the patient s movement during these episodes and machine learning algorithms can potentially classify this data the combination therefore holds the potential to detect fog in real time in this work i present a simple d convolutional neural network that was trained to detect fog events in accelerometer data model performance was assessed by measuring the success of the model to discriminate normal movement from fog episodes and resulted in a mean average precision of on the private leaderboard on kaggle ultimately the model ranked out of teams in the parkinson s freezing of gait prediction competition the results underscore the potential of deep learning based solutions in advancing the field of fog detection contributing to improved interventions and management strategies for pd patients keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb tranfer learning of semantic segmentation methods for identifying buried archaeological structures on lidar data authors paolo soleni wouter b verschoof van der vaart žiga kokalj arianna traviglia marco fiorucci subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract when applying deep learning to remote sensing data in archaeological research a notable obstacle is the limited availability of suitable datasets for training models the application of transfer learning is frequently employed to mitigate this drawback however there is still a need to explore its effectiveness when applied across different archaeological datasets this paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two lidar datasets the experimental results indicate that transfer learning based approaches in archaeology can lead to performance improvements although a systematic enhancement has not yet been observed we provide specific insights about the validity of such techniques that can serve as a baseline for future works keyword isp general purpose multimodal transformer meets remote sensing semantic segmentation authors nhi kieu kien nguyen sridha sridharan clinton fookes subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the advent of high resolution multispectral hyperspectral sensors lidar dsm digital surface model information and many others has provided us with an unprecedented wealth of data for earth observation multimodal ai seeks to exploit those complementary data sources particularly for complex tasks like semantic segmentation while specialized architectures have been developed they are highly complicated via significant effort in model design and require considerable re engineering whenever a new modality emerges recent trends in general purpose multimodal networks have shown great potential to achieve state of the art performance across multiple multimodal tasks with one unified architecture in this work we investigate the performance of perceiverio one in the general purpose multimodal family in the remote sensing semantic segmentation domain our experiments reveal that this ostensibly universal network struggles with object scale variation in remote sensing images and fails to detect the presence of cars from a top down view to address these issues even with extreme class imbalance issues we propose a spatial and volumetric learning component specifically we design a unet inspired module that employs convolution to encode vital local information and learn cross modal features simultaneously while reducing network computational burden via the cross attention mechanism of perceiverio the effectiveness of the proposed component is validated through extensive experiments comparing it with other methods such as convolution and dual local module ie the combination of and inspired by unetformer the proposed method achieves competitive results with specialized architectures like unetformer and swinunet showing its potential to minimize network architecture engineering with a minimal compromise on the performance unsupervised hyperspectral and multispectral images fusion based on the cycle consistency authors shuaikai shi lijun zhang yoann altmann jie chen subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract hyperspectral images hsi with abundant spectral information reflected materials property usually perform low spatial resolution due to the hardware limits meanwhile multispectral images msi e g rgb images have a high spatial resolution but deficient spectral signatures hyperspectral and multispectral image fusion can be cost effective and efficient for acquiring both high spatial resolution and high spectral resolution images many of the conventional hsi and msi fusion algorithms rely on known spatial degradation parameters i e point spread function spectral degradation parameters spectral response function or both of them another class of deep learning based models relies on the ground truth of high spatial resolution hsi and needs large amounts of paired training images when working in a supervised manner both of these models are limited in practical fusion scenarios in this paper we propose an unsupervised hsi and msi fusion model based on the cycle consistency called cycfusion the cycfusion learns the domain transformation between low spatial resolution hsi lrhsi and high spatial resolution msi hrmsi and the desired high spatial resolution hsi hrhsi are considered to be intermediate feature maps in the transformation networks the cycfusion can be trained with the objective functions of marginal matching in single transform and cycle consistency in double transforms moreover the estimated psf and srf are embedded in the model as the pre training weights which further enhances the practicality of our proposed model experiments conducted on several datasets show that our proposed model outperforms all compared unsupervised fusion methods the codes of this paper will be available at this address https github com shuaikaishi cycfusion for reproducibility keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw all in one exploring unified vision language tracking with multi modal alignment authors chunhui zhang xin sun li liu yiqian yang qiong liu xi zhou yanfeng wang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract current mainstream vision language vl tracking framework consists of three parts ie a visual feature extractor a language feature extractor and a fusion model to pursue better performance a natural modus operandi for vl tracking is employing customized and heavier unimodal encoders and multi modal fusion models albeit effective existing vl trackers separate feature extraction and feature integration resulting in extracted features that lack semantic guidance and have limited target aware capability in complex scenarios eg similar distractors and extreme illumination in this work inspired by the recent success of exploring foundation models with unified architecture for both natural language and computer vision tasks we propose an all in one framework which learns joint feature extraction and interaction by adopting a unified transformer backbone specifically we mix raw vision and language signals to generate language injected vision tokens which we then concatenate before feeding into the unified backbone architecture this approach achieves feature integration in a unified backbone removing the need for carefully designed fusion modules and resulting in a more effective and efficient vl tracking framework to further improve the learning efficiency we introduce a multi modal alignment module based on cross modal and intra modal contrastive objectives providing more reasonable representations for the unified all in one transformer backbone extensive experiments on five benchmarks ie l lasot lasot rm ext and webuav demonstrate the superiority of the proposed tracker against existing state of the arts on vl tracking codes will be made publicly available tranfer learning of semantic segmentation methods for identifying buried archaeological structures on lidar data authors paolo soleni wouter b verschoof van der vaart žiga kokalj arianna traviglia marco fiorucci subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract when applying deep learning to remote sensing data in archaeological research a notable obstacle is the limited availability of suitable datasets for training models the application of transfer learning is frequently employed to mitigate this drawback however there is still a need to explore its effectiveness when applied across different archaeological datasets this paper compares the performance of various transfer learning configurations using two semantic segmentation deep neural networks on two lidar datasets the experimental results indicate that transfer learning based approaches in archaeology can lead to performance improvements although a systematic enhancement has not yet been observed we provide specific insights about the validity of such techniques that can serve as a baseline for future works language free compositional action generation via decoupling refinement authors xiao liu guangyi chen yansong tang guangrun wang ser nam lim subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract composing simple elements into complex concepts is crucial yet challenging especially for action generation existing methods largely rely on extensive neural language annotations to discern composable latent semantics a process that is often costly and labor intensive in this study we introduce a novel framework to generate compositional actions without reliance on language auxiliaries our approach consists of three main components action coupling conditional action generation and decoupling refinement action coupling utilizes an energy model to extract the attention masks of each sub action subsequently integrating two actions using these attentions to generate pseudo training examples then we employ a conditional generative model cvae to learn a latent space facilitating the diverse generation finally we propose decoupling refinement which leverages a self supervised pre trained model mae to ensure semantic consistency between the sub actions and compositional actions this refinement process involves rendering generated actions into space decoupling these images into two sub segments using the mae model to restore the complete image from sub segments and constraining the recovered images to match images rendered from raw sub actions due to the lack of existing datasets containing both sub actions and compositional actions we created two new datasets named humanact c and uestc c and present a corresponding evaluation metric both qualitative and quantitative assessments are conducted to show our efficacy keyword raw image there is no result
| 1
|
2,392
| 5,190,341,474
|
IssuesEvent
|
2017-01-21 07:34:30
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
process: misaligned node --help text
|
good first contribution process
|
* **Version**: master
* **Platform**: n/a
* **Subsystem**: process
The alignment of the argument descriptions in the `node --help` text is off. Specifically the `--inspect` and `--inspect-brk` options do not match up. Because the full `--inspect-brk` argument name is the longest, all of the other arguments' descriptions may need to be indented a couple of spaces further to accommodate this longer argument name.
|
1.0
|
process: misaligned node --help text - * **Version**: master
* **Platform**: n/a
* **Subsystem**: process
The alignment of the argument descriptions in the `node --help` text is off. Specifically the `--inspect` and `--inspect-brk` options do not match up. Because the full `--inspect-brk` argument name is the longest, all of the other arguments' descriptions may need to be indented a couple of spaces further to accommodate this longer argument name.
|
process
|
process misaligned node help text version master platform n a subsystem process the alignment of the argument descriptions in the node help text is off specifically the inspect and inspect brk options do not match up because the full inspect brk argument name is the longest all of the other arguments descriptions may need to be indented a couple of spaces further to accommodate this longer argument name
| 1
|
10,671
| 13,458,564,998
|
IssuesEvent
|
2020-09-09 10:51:08
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
opened
|
[FALSE-POSITIVE?] *ionicframework.com
|
whitelisting process
|
**Domains or links**
Please list any domains and links listed here which you believe are a false positive.
*ionicframework.com
**More Information**
How did you discover your web site or domain was listed here?
- HostsVN - unknown reason
- https://hosts.ubuntu101.co.za/app/contents/ says also in https://raw.githubusercontent.com/anudeepND/blacklist/master/adservers.txt but can't find it here @funilrys
**Have you requested removal from other sources?**
No because I cannot speak vietnamese
**Additional context**
Currently I see no reason why the project site of the ionic framework was blocked.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
1.0
|
[FALSE-POSITIVE?] *ionicframework.com - **Domains or links**
Please list any domains and links listed here which you believe are a false positive.
*ionicframework.com
**More Information**
How did you discover your web site or domain was listed here?
- HostsVN - unknown reason
- https://hosts.ubuntu101.co.za/app/contents/ says also in https://raw.githubusercontent.com/anudeepND/blacklist/master/adservers.txt but can't find it here @funilrys
**Have you requested removal from other sources?**
No because I cannot speak vietnamese
**Additional context**
Currently I see no reason why the project site of the ionic framework was blocked.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
process
|
ionicframework com domains or links please list any domains and links listed here which you believe are a false positive ionicframework com more information how did you discover your web site or domain was listed here hostsvn unknown reason says also in but can t find it here funilrys have you requested removal from other sources no because i cannot speak vietnamese additional context currently i see no reason why the project site of the ionic framework was blocked exclamation we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
| 1
|
1,630
| 4,241,046,700
|
IssuesEvent
|
2016-07-06 15:13:03
|
gratipay/inside.gratipay.com
|
https://api.github.com/repos/gratipay/inside.gratipay.com
|
closed
|
send whit537 to OS//OS
|
Events Governance & Process Marketing Partnerships
|
[Open Source, Open Society](http://opensourceopensociety.com/) is a conference in New Zealand in August.
> Welcome to the world where being open is better for business, technology and democracy.
Going there would further our relationship-building in the new economy movement, following on from Platform Cooperativism (#384) and OuiShare Fest (#314). Of particular interest, the conference is co-hosted by Enspiral, a group that is both a significant player in the new economy movement, and also highly geared towards face-to-face relationships. Going to OS//OS would help a) deepen our relationship with Enspiral, b) potentially clarify #421, and c) contribute to #72. @ntnsndr, recently returned from New Zealand, suggested under https://github.com/gratipay/inside.gratipay.com/issues/72#issuecomment-229196904 that Alanna Krause and (Enspiral ex-founder) Joshua Vial would be particularly good connections for us to make. I've [confirmed](https://twitter.com/whit537/status/748185210299133953) that both are planning to be at OS//OS.
Bonus points: our [registrar](https://iwantmyname.com/) is a [sponsor](http://opensourceopensociety.com/partners.html)! :o)
|
1.0
|
send whit537 to OS//OS - [Open Source, Open Society](http://opensourceopensociety.com/) is a conference in New Zealand in August.
> Welcome to the world where being open is better for business, technology and democracy.
Going there would further our relationship-building in the new economy movement, following on from Platform Cooperativism (#384) and OuiShare Fest (#314). Of particular interest, the conference is co-hosted by Enspiral, a group that is both a significant player in the new economy movement, and also highly geared towards face-to-face relationships. Going to OS//OS would help a) deepen our relationship with Enspiral, b) potentially clarify #421, and c) contribute to #72. @ntnsndr, recently returned from New Zealand, suggested under https://github.com/gratipay/inside.gratipay.com/issues/72#issuecomment-229196904 that Alanna Krause and (Enspiral ex-founder) Joshua Vial would be particularly good connections for us to make. I've [confirmed](https://twitter.com/whit537/status/748185210299133953) that both are planning to be at OS//OS.
Bonus points: our [registrar](https://iwantmyname.com/) is a [sponsor](http://opensourceopensociety.com/partners.html)! :o)
|
process
|
send to os os is a conference in new zealand in august welcome to the world where being open is better for business technology and democracy going there would further our relationship building in the new economy movement following on from platform cooperativism and ouishare fest of particular interest the conference is co hosted by enspiral a group that is both a significant player in the new economy movement and also highly geared towards face to face relationships going to os os would help a deepen our relationship with enspiral b potentially clarify and c contribute to ntnsndr recently returned from new zealand suggested under that alanna krause and enspiral ex founder joshua vial would be particularly good connections for us to make i ve that both are planning to be at os os bonus points our is a o
| 1
|
114,390
| 9,702,889,020
|
IssuesEvent
|
2019-05-27 09:52:15
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
opened
|
The percentage is changed from 7% to 100% directly when uploading a page blob using AzCopy
|
:gear: azcopy 🧪 testing
|
**Storage Explorer Version:** 1.8.1_20190524.5
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Make sure AzCopy is enabled.
2. Expand one normal account -> 'Blob Containers'.
3. Open one blob container then upload a large blob file(Like 30G) as page blob.
4. Check the percentage from uploading activities.
**Expect Experience:**
The percentage is getting bigger and bigger.
**Actual Experience:**
The percentage keep 7% a long time, then upload activities finishes directly.

**More info:**
1. This issue doesn't reproduce when uploading a block blob using AzCopy.
2. This issue doesn't reproduce when uploading a page/block blob without AzCopy.
|
1.0
|
The percentage is changed from 7% to 100% directly when uploading a page blob using AzCopy - **Storage Explorer Version:** 1.8.1_20190524.5
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Make sure AzCopy is enabled.
2. Expand one normal account -> 'Blob Containers'.
3. Open one blob container then upload a large blob file(Like 30G) as page blob.
4. Check the percentage from uploading activities.
**Expect Experience:**
The percentage is getting bigger and bigger.
**Actual Experience:**
The percentage keep 7% a long time, then upload activities finishes directly.

**More info:**
1. This issue doesn't reproduce when uploading a block blob using AzCopy.
2. This issue doesn't reproduce when uploading a page/block blob without AzCopy.
|
non_process
|
the percentage is changed from to directly when uploading a page blob using azcopy storage explorer version platform os linux ubuntu macos high sierra windows architecture regression from not a regression steps to reproduce make sure azcopy is enabled expand one normal account blob containers open one blob container then upload a large blob file like as page blob check the percentage from uploading activities expect experience the percentage is getting bigger and bigger actual experience the percentage keep a long time then upload activities finishes directly more info this issue doesn t reproduce when uploading a block blob using azcopy this issue doesn t reproduce when uploading a page block blob without azcopy
| 0
|
34,926
| 14,544,788,502
|
IssuesEvent
|
2020-12-15 18:41:22
|
NetsBlox/NetsBlox
|
https://api.github.com/repos/NetsBlox/NetsBlox
|
closed
|
Connect to custom CoreNLP server
|
enhancement services
|
The CoreNLP server should support connecting to custom URLs (perhaps via an environment variable). This would allow us to update the docker-compose to include a CoreNLP instance (using an image like [this one](https://hub.docker.com/r/graham3333/corenlp-complete)).
|
1.0
|
Connect to custom CoreNLP server - The CoreNLP server should support connecting to custom URLs (perhaps via an environment variable). This would allow us to update the docker-compose to include a CoreNLP instance (using an image like [this one](https://hub.docker.com/r/graham3333/corenlp-complete)).
|
non_process
|
connect to custom corenlp server the corenlp server should support connecting to custom urls perhaps via an environment variable this would allow us to update the docker compose to include a corenlp instance using an image like
| 0
|
21,644
| 3,738,864,202
|
IssuesEvent
|
2016-03-09 00:55:28
|
department-of-veterans-affairs/roadrunner
|
https://api.github.com/repos/department-of-veterans-affairs/roadrunner
|
closed
|
Grid Template
|
2-in-progress design front end
|
Exactly like USWDS https://playbook.cio.gov/designstandards/grids/
- Grid
- Grid Code
- Grid Example
- Grid Example Code
- Documentation
Only difference is our Header text is in Roboto Slab; #0071BC
Description Text: This 12-column, responsive grid provides structure for website content.
<img width="969" alt="screen shot 2016-03-03 at 9 19 13 am" src="https://cloud.githubusercontent.com/assets/13420618/13502754/1eb90502-e121-11e5-8b65-81b8c857442b.png">
|
1.0
|
Grid Template - Exactly like USWDS https://playbook.cio.gov/designstandards/grids/
- Grid
- Grid Code
- Grid Example
- Grid Example Code
- Documentation
Only difference is our Header text is in Roboto Slab; #0071BC
Description Text: This 12-column, responsive grid provides structure for website content.
<img width="969" alt="screen shot 2016-03-03 at 9 19 13 am" src="https://cloud.githubusercontent.com/assets/13420618/13502754/1eb90502-e121-11e5-8b65-81b8c857442b.png">
|
non_process
|
grid template exactly like uswds grid grid code grid example grid example code documentation only difference is our header text is in roboto slab description text this column responsive grid provides structure for website content img width alt screen shot at am src
| 0
|
5,915
| 7,404,458,819
|
IssuesEvent
|
2018-03-20 04:55:24
|
vmware/harbor
|
https://api.github.com/repos/vmware/harbor
|
opened
|
Support status hook function in the new job service.
|
area/job-services
|
Use the job status hook to report and update job status.
|
1.0
|
Support status hook function in the new job service. - Use the job status hook to report and update job status.
|
non_process
|
support status hook function in the new job service use the job status hook to report and update job status
| 0
|
23,692
| 6,474,865,914
|
IssuesEvent
|
2017-08-17 19:03:28
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
Bad interaction between constant prop and CSE
|
area-CodeGen enhancement optimization tenet-performance
|
Exploring the invariant hoisting code I found that it doesn't work with the assignments.
In optimizer.cpp there's a code:
```
if (treeIsInvariant)
{
// Tree must be a suitable CSE candidate for us to be able to hoist it.
treeIsHoistable = optIsCSEcandidate(tree);
```
optIsCSEcandidate is false for assignments.
Though it won't work with the assignment the analysis will detect that the the expression subtree of the assignment is invariant and will actually clone the expression to the prehead block of the loop.
We'll get a tree looking e.g. like this:
```
------------ BB04 [029..???), preds={BB01} succs={BB02}
***** BB04, stmt 8 (top level)
( 5, 5) [000098] ------------- * stmtExpr void (top level) (IL ???... ???)
N006 ( 0, 0) [000096] ------------- | /--* nop void
N007 ( 5, 5) [000097] ------------- \--* comma void
N004 ( 1, 1) [000095] -------H----- | /--* lclVar int V04 loc3 u:3 <l:$46, c:$1c1>
N005 ( 5, 5) [000091] -------H----- \--* + int <l:$4a, c:$1c6>
N002 ( 1, 1) [000094] -------H----- | /--* lclVar int V03 loc2 u:3 <l:$44, c:$1c0>
N003 ( 3, 3) [000092] -------H----- \--* + int <l:$49, c:$1c5>
N001 ( 1, 1) [000093] -------H----- \--* lclVar int V02 loc1 u:3 <l:$42, c:$101>
------------ BB02 [029..043) -> BB02 (cond), preds={BB02,BB04} succs={BB03,BB02}
```
The result of such a tree is not used and on the code generator no code is produced for such a tree.
/cc @BruceForstall @brucehoult @briansull @Dmitri-Botcharnikov
|
1.0
|
Bad interaction between constant prop and CSE - Exploring the invariant hoisting code I found that it doesn't work with the assignments.
In optimizer.cpp there's a code:
```
if (treeIsInvariant)
{
// Tree must be a suitable CSE candidate for us to be able to hoist it.
treeIsHoistable = optIsCSEcandidate(tree);
```
optIsCSEcandidate is false for assignments.
Though it won't work with the assignment the analysis will detect that the the expression subtree of the assignment is invariant and will actually clone the expression to the prehead block of the loop.
We'll get a tree looking e.g. like this:
```
------------ BB04 [029..???), preds={BB01} succs={BB02}
***** BB04, stmt 8 (top level)
( 5, 5) [000098] ------------- * stmtExpr void (top level) (IL ???... ???)
N006 ( 0, 0) [000096] ------------- | /--* nop void
N007 ( 5, 5) [000097] ------------- \--* comma void
N004 ( 1, 1) [000095] -------H----- | /--* lclVar int V04 loc3 u:3 <l:$46, c:$1c1>
N005 ( 5, 5) [000091] -------H----- \--* + int <l:$4a, c:$1c6>
N002 ( 1, 1) [000094] -------H----- | /--* lclVar int V03 loc2 u:3 <l:$44, c:$1c0>
N003 ( 3, 3) [000092] -------H----- \--* + int <l:$49, c:$1c5>
N001 ( 1, 1) [000093] -------H----- \--* lclVar int V02 loc1 u:3 <l:$42, c:$101>
------------ BB02 [029..043) -> BB02 (cond), preds={BB02,BB04} succs={BB03,BB02}
```
The result of such a tree is not used and on the code generator no code is produced for such a tree.
/cc @BruceForstall @brucehoult @briansull @Dmitri-Botcharnikov
|
non_process
|
bad interaction between constant prop and cse exploring the invariant hoisting code i found that it doesn t work with the assignments in optimizer cpp there s a code if treeisinvariant tree must be a suitable cse candidate for us to be able to hoist it treeishoistable optiscsecandidate tree optiscsecandidate is false for assignments though it won t work with the assignment the analysis will detect that the the expression subtree of the assignment is invariant and will actually clone the expression to the prehead block of the loop we ll get a tree looking e g like this preds succs stmt top level stmtexpr void top level il nop void comma void h lclvar int u h int h lclvar int u h int h lclvar int u cond preds succs the result of such a tree is not used and on the code generator no code is produced for such a tree cc bruceforstall brucehoult briansull dmitri botcharnikov
| 0
|
6,078
| 8,923,531,759
|
IssuesEvent
|
2019-01-21 15:53:52
|
enKryptIO/ethvm
|
https://api.github.com/repos/enKryptIO/ethvm
|
closed
|
Add a concept of chain config to allow for processing adjustments based on block number
|
bug milestone:1 priority:high project:processing
|
Some fields start behaving differently and certain constraints are different based on the block number. Currently it appears that tx success is incorrect due to such a difference.
We need to introduce a concept of chain config which will contain various feature flags to allow us to process things differently.
|
1.0
|
Add a concept of chain config to allow for processing adjustments based on block number - Some fields start behaving differently and certain constraints are different based on the block number. Currently it appears that tx success is incorrect due to such a difference.
We need to introduce a concept of chain config which will contain various feature flags to allow us to process things differently.
|
process
|
add a concept of chain config to allow for processing adjustments based on block number some fields start behaving differently and certain constraints are different based on the block number currently it appears that tx success is incorrect due to such a difference we need to introduce a concept of chain config which will contain various feature flags to allow us to process things differently
| 1
|
36,712
| 4,755,891,063
|
IssuesEvent
|
2016-10-24 12:24:00
|
Automattic/jetpack
|
https://api.github.com/repos/Automattic/jetpack
|
opened
|
Admin Page: add link to wp.com/me in Connection settings
|
Admin Page [Status] Needs Design Review [Type] Enhancement
|
> In the JetPack settings page, where we can see the status of “Connection Settings”, you need to include a link or button link to go to our WordPress Management Page, and this will open in a new window.
Suggested here:
https://wordpress.org/support/topic/needs-addition-link-to-wordpress-management-account/
cc @MichaelArestad
|
1.0
|
Admin Page: add link to wp.com/me in Connection settings - > In the JetPack settings page, where we can see the status of “Connection Settings”, you need to include a link or button link to go to our WordPress Management Page, and this will open in a new window.
Suggested here:
https://wordpress.org/support/topic/needs-addition-link-to-wordpress-management-account/
cc @MichaelArestad
|
non_process
|
admin page add link to wp com me in connection settings in the jetpack settings page where we can see the status of “connection settings” you need to include a link or button link to go to our wordpress management page and this will open in a new window suggested here cc michaelarestad
| 0
|
12,073
| 4,351,568,584
|
IssuesEvent
|
2016-07-31 23:04:54
|
colinhect/hect
|
https://api.github.com/repos/colinhect/hect
|
closed
|
Formalize Scene types similar with Systems and Components
|
api code-quality
|
Add a SceneRegistery class. Have specific scene types derive from Scene and register with SceneRegistery. The scene class will define which systems it has and defines its own tick() function (removing the need to describe system update order in data). The scene data itself will dictate its scene type. Scene base types should still work as-is.
|
1.0
|
Formalize Scene types similar with Systems and Components - Add a SceneRegistery class. Have specific scene types derive from Scene and register with SceneRegistery. The scene class will define which systems it has and defines its own tick() function (removing the need to describe system update order in data). The scene data itself will dictate its scene type. Scene base types should still work as-is.
|
non_process
|
formalize scene types similar with systems and components add a sceneregistery class have specific scene types derive from scene and register with sceneregistery the scene class will define which systems it has and defines its own tick function removing the need to describe system update order in data the scene data itself will dictate its scene type scene base types should still work as is
| 0
|
11,187
| 13,957,697,880
|
IssuesEvent
|
2020-10-24 08:12:15
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
IE: INSPIRE Geoportal Harvesting Checker
|
Geoportal Harvesting process IE - Ireland
|
From: Ben Scott [benscott@esri-ireland.ie]
Sent: 01 October 2018 18:49
To: JRC INSPIRE SUPPORT
Subject: INSPIRE Geoportal Harvesting Checker
Hi – I am managing a project on behalf of Ordnance Survey Ireland to look at options to implement a new INSPIRE portal for Ireland. We are trying to determine the EU harvesting requirements for metadata.
Is a CSW the only format that can be used to make a selection of metadata available for harvest by the EU or can it also be provided in other formats, for example in a Web Accessible Folder?
Regards,
Ben Scott
|
1.0
|
IE: INSPIRE Geoportal Harvesting Checker - From: Ben Scott [benscott@esri-ireland.ie]
Sent: 01 October 2018 18:49
To: JRC INSPIRE SUPPORT
Subject: INSPIRE Geoportal Harvesting Checker
Hi – I am managing a project on behalf of Ordnance Survey Ireland to look at options to implement a new INSPIRE portal for Ireland. We are trying to determine the EU harvesting requirements for metadata.
Is a CSW the only format that can be used to make a selection of metadata available for harvest by the EU or can it also be provided in other formats, for example in a Web Accessible Folder?
Regards,
Ben Scott
|
process
|
ie inspire geoportal harvesting checker from ben scott sent october to jrc inspire support subject inspire geoportal harvesting checker hi ndash i am managing a project on behalf of ordnance survey ireland to look at options to implement a new inspire portal for ireland we are trying to determine the eu harvesting requirements for metadata is a csw the only format that can be used to make a selection of metadata available for harvest by the eu or can it also be provided in other formats for example in a web accessible folder regards ben scott
| 1
|
36,352
| 7,895,356,436
|
IssuesEvent
|
2018-06-29 02:44:57
|
niltonvolpato/python-progressbar
|
https://api.github.com/repos/niltonvolpato/python-progressbar
|
closed
|
Your own examples don't run
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. virtualenv .
2. source bin/activate
3. pip install progressbar
4. wget http://python-progressbar.googlecode.com/hg/examples.py
5. python examples.py
What is the expected output? What do you see instead?
The expected output is 19 different examples of progress bars.
The actual output is:
Traceback (most recent call last):
File "examples.py", line 7, in <module>
from progressbar import AnimatedMarker, Bar, BouncingBar, Counter, ETA, \
ImportError: cannot import name AnimatedMarker
What version of the product are you using? On what operating system?
pip installs version 2.2. The operating system is linux mint (an ubuntu
derivative).
Please provide any additional information below.
My initial attempts to use this library were as simple as:
python -c 'import time
from progressbar import ProgressBar
progress = ProgressBar()
for i in progress(range(80)):
time.sleep(0.01)'
(based on example 19). This does not work:
Traceback (most recent call last):
File "<string>", line 5, in <module>
TypeError: 'ProgressBar' object is not callable
```
Original issue reported on code.google.com by `ma...@gapps.semantico.com` on 25 Feb 2014 at 10:48
|
1.0
|
Your own examples don't run - ```
What steps will reproduce the problem?
1. virtualenv .
2. source bin/activate
3. pip install progressbar
4. wget http://python-progressbar.googlecode.com/hg/examples.py
5. python examples.py
What is the expected output? What do you see instead?
The expected output is 19 different examples of progress bars.
The actual output is:
Traceback (most recent call last):
File "examples.py", line 7, in <module>
from progressbar import AnimatedMarker, Bar, BouncingBar, Counter, ETA, \
ImportError: cannot import name AnimatedMarker
What version of the product are you using? On what operating system?
pip installs version 2.2. The operating system is linux mint (an ubuntu
derivative).
Please provide any additional information below.
My initial attempts to use this library were as simple as:
python -c 'import time
from progressbar import ProgressBar
progress = ProgressBar()
for i in progress(range(80)):
time.sleep(0.01)'
(based on example 19). This does not work:
Traceback (most recent call last):
File "<string>", line 5, in <module>
TypeError: 'ProgressBar' object is not callable
```
Original issue reported on code.google.com by `ma...@gapps.semantico.com` on 25 Feb 2014 at 10:48
|
non_process
|
your own examples don t run what steps will reproduce the problem virtualenv source bin activate pip install progressbar wget python examples py what is the expected output what do you see instead the expected output is different examples of progress bars the actual output is traceback most recent call last file examples py line in from progressbar import animatedmarker bar bouncingbar counter eta importerror cannot import name animatedmarker what version of the product are you using on what operating system pip installs version the operating system is linux mint an ubuntu derivative please provide any additional information below my initial attempts to use this library were as simple as python c import time from progressbar import progressbar progress progressbar for i in progress range time sleep based on example this does not work traceback most recent call last file line in typeerror progressbar object is not callable original issue reported on code google com by ma gapps semantico com on feb at
| 0
|
217,851
| 16,740,819,327
|
IssuesEvent
|
2021-06-11 09:33:11
|
exasol/generic-virtual-schema
|
https://api.github.com/repos/exasol/generic-virtual-schema
|
closed
|
Add the documentation
|
documentation source:exasol
|
## Problem
The generic dialect doesn't have a user guide.
We want to add a user guide with a small tutorial how to user it.
|
1.0
|
Add the documentation - ## Problem
The generic dialect doesn't have a user guide.
We want to add a user guide with a small tutorial how to user it.
|
non_process
|
add the documentation problem the generic dialect doesn t have a user guide we want to add a user guide with a small tutorial how to user it
| 0
|
14,777
| 18,051,611,625
|
IssuesEvent
|
2021-09-19 20:58:50
|
nanoframework/Home
|
https://api.github.com/repos/nanoframework/Home
|
reopened
|
Support for generics <T>
|
Status: In progress Type: Feature request Area: Interpreter Area: Visual Studio extension Area: Debugger-Library :pushpin: pinned Area: Metadata Processor
|
**nanoFramework area:** (CLR)
We are actively working on adding support for generics.
- [x] CLR
- [x] Type System
- [x] Interpreter
- [x] Debugger Library
- [x] Metadata processor
- [ ] Support in Visual Studio debugger engine
<!-- feature-request-tag DO NOT REMOVE -->
|
1.0
|
Support for generics <T> - **nanoFramework area:** (CLR)
We are actively working on adding support for generics.
- [x] CLR
- [x] Type System
- [x] Interpreter
- [x] Debugger Library
- [x] Metadata processor
- [ ] Support in Visual Studio debugger engine
<!-- feature-request-tag DO NOT REMOVE -->
|
process
|
support for generics nanoframework area clr we are actively working on adding support for generics clr type system interpreter debugger library metadata processor support in visual studio debugger engine
| 1
|
811,070
| 30,273,527,253
|
IssuesEvent
|
2023-07-07 17:24:18
|
recro/ats-issues
|
https://api.github.com/repos/recro/ats-issues
|
closed
|
Cannot use filter options in grids
|
bug area/candidate priority/critical area/application area/job
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
I am unable to use the filters for fields where there are multiple selections (contract, application status, candidate owner, clearance, etc.). when the 3 lines are selected, nothing pops up.
### Expected behavior
Need to be able to use the filters to sort the data
### Steps to reproduce
_No response_
### Exceptions or error messages (if any)
_No response_
### Anything else?
_No response_
|
1.0
|
Cannot use filter options in grids - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
I am unable to use the filters for fields where there are multiple selections (contract, application status, candidate owner, clearance, etc.). when the 3 lines are selected, nothing pops up.
### Expected behavior
Need to be able to use the filters to sort the data
### Steps to reproduce
_No response_
### Exceptions or error messages (if any)
_No response_
### Anything else?
_No response_
|
non_process
|
cannot use filter options in grids is there an existing issue for this i have searched the existing issues describe the bug i am unable to use the filters for fields where there are multiple selections contract application status candidate owner clearance etc when the lines are selected nothing pops up expected behavior need to be able to use the filters to sort the data steps to reproduce no response exceptions or error messages if any no response anything else no response
| 0
|
92,541
| 3,872,205,175
|
IssuesEvent
|
2016-04-11 13:07:36
|
sonejostudios/superboucle
|
https://api.github.com/repos/sonejostudios/superboucle
|
closed
|
Load songs on superboucle start
|
enhancement priority: high
|
2 cases
* production: remember the song that was loaded last sb session
* performance: start sb with 1 positional argument: songpath
|
1.0
|
Load songs on superboucle start - 2 cases
* production: remember the song that was loaded last sb session
* performance: start sb with 1 positional argument: songpath
|
non_process
|
load songs on superboucle start cases production remember the song that was loaded last sb session performance start sb with positional argument songpath
| 0
|
10,432
| 13,219,894,055
|
IssuesEvent
|
2020-08-17 11:22:09
|
bisq-network/proposals
|
https://api.github.com/repos/bisq-network/proposals
|
closed
|
Trading protocol change: release of funds in 2of2 multisig to be signed by seller in first place.
|
a:proposal re:processes was:stalled
|
With the objective to prevent seller from performing [future trades](https://github.com/bisq-network/proposals/issues/220#issuecomment-629582705), I propose to change the trading protocol to make seller be the first to sign the 2of2 multisig and let buyer sign the normal payout (buyer gets trade funds, both parts get their deposits back) at the end of the trade settlement.
This is how the trade protocol changes:
1. Buyer deposits +15% deposit and buyer deposits trade funds and 15% deposit.
2. Buyer sends fiat or altcoin payment to seller, press `payment sent `button ~~and provides the first signature of the 2of2 multisig tx.~~
3. Seller clicks `payment received`, ~~signing and publishing the multisig tx~~ and provides the first signature of the 2of2 multisig tx.
4. Buyer signs and publishes the multisig tx, releasing deposits and trade funds.
With the current protocol, when buyer pushes the `payment sent` button, provides the first signature of the 2of2 multisig. Even if buyer starts a dispute, seller can just sign the second signature, [overriding mediator's suggestion](https://github.com/bisq-network/bisq/issues/4162).
This proposal adds an extra step (number 4) that makes future trading riskier as it does not allow sellers to omit mediation, so in case he does not push the `payment received button` on time, he could loose the security deposit.
If, as proposed, buyer doesn't provide any signature when clicks payment sent, he will be able to open a dispute that will have effect. Buyer has more to loose in case of mediation after payment has been sent, so if he's the last to sign, we should expect buyer will release funds according to protocol.
|
1.0
|
Trading protocol change: release of funds in 2of2 multisig to be signed by seller in first place. - With the objective to prevent seller from performing [future trades](https://github.com/bisq-network/proposals/issues/220#issuecomment-629582705), I propose to change the trading protocol to make seller be the first to sign the 2of2 multisig and let buyer sign the normal payout (buyer gets trade funds, both parts get their deposits back) at the end of the trade settlement.
This is how the trade protocol changes:
1. Buyer deposits +15% deposit and buyer deposits trade funds and 15% deposit.
2. Buyer sends fiat or altcoin payment to seller, press `payment sent `button ~~and provides the first signature of the 2of2 multisig tx.~~
3. Seller clicks `payment received`, ~~signing and publishing the multisig tx~~ and provides the first signature of the 2of2 multisig tx.
4. Buyer signs and publishes the multisig tx, releasing deposits and trade funds.
With the current protocol, when buyer pushes the `payment sent` button, provides the first signature of the 2of2 multisig. Even if buyer starts a dispute, seller can just sign the second signature, [overriding mediator's suggestion](https://github.com/bisq-network/bisq/issues/4162).
This proposal adds an extra step (number 4) that makes future trading riskier as it does not allow sellers to omit mediation, so in case he does not push the `payment received button` on time, he could loose the security deposit.
If, as proposed, buyer doesn't provide any signature when clicks payment sent, he will be able to open a dispute that will have effect. Buyer has more to loose in case of mediation after payment has been sent, so if he's the last to sign, we should expect buyer will release funds according to protocol.
|
process
|
trading protocol change release of funds in multisig to be signed by seller in first place with the objective to prevent seller from performing i propose to change the trading protocol to make seller be the first to sign the multisig and let buyer sign the normal payout buyer gets trade funds both parts get their deposits back at the end of the trade settlement this is how the trade protocol changes buyer deposits deposit and buyer deposits trade funds and deposit buyer sends fiat or altcoin payment to seller press payment sent button and provides the first signature of the multisig tx seller clicks payment received signing and publishing the multisig tx and provides the first signature of the multisig tx buyer signs and publishes the multisig tx releasing deposits and trade funds with the current protocol when buyer pushes the payment sent button provides the first signature of the multisig even if buyer starts a dispute seller can just sign the second signature this proposal adds an extra step number that makes future trading riskier as it does not allow sellers to omit mediation so in case he does not push the payment received button on time he could loose the security deposit if as proposed buyer doesn t provide any signature when clicks payment sent he will be able to open a dispute that will have effect buyer has more to loose in case of mediation after payment has been sent so if he s the last to sign we should expect buyer will release funds according to protocol
| 1
|
3,934
| 6,849,958,427
|
IssuesEvent
|
2017-11-14 00:28:45
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
process.send called in a loop in a forked child process eats up heap.
|
child_process process question
|
* **Version**: 8.9.1
* **Platform**: Windows 10 64 bit
* **Subsystem**: Child Process
<!-- Enter your issue details below this comment. -->
Similar-sounding issue [here](https://github.com/nodejs/node/issues/15651) but event based / not in a loop.
I've managed to reduce the used heap in a test by ~400K by forking with the` --expose-gc` option and calling `global.gc` after calling `process.send`.
parent.js ;
```
"use strict" ;
var forkOptions = {
execArgv:['--expose-gc']
};
const cp = require('child_process');
const child = cp.fork(`./child.js`,[],forkOptions);
child.on("message",function(m){
//console.log(m);
});
```
child.js ;
```
// invoke with --expose-gc
"use strict" ;
var fs = require('fs');
const loopsToWatch = 100 ;
const threshold = 32 ; // bytes
function underTest(){
// let newExtractDirFiles = fs.readdirSync("./"); // threshold = 32
process.send("yada");
}
let newMem = process.memoryUsage().heapUsed ;
let increases = 0 , decreases = 0 , maxIncrease = 0 , maxUsed = 0 ;
let mem = process.memoryUsage().heapUsed ;
let intialMem = mem ;
for(let i = 0 ; i < loopsToWatch ; i++){
underTest();
global.gc() ; //requires --expose-gc flag
newMem = process.memoryUsage().heapUsed ;
if(newMem > maxUsed){
maxUsed = newMem ;
}
let diff = newMem - mem ;
if(diff > maxIncrease){
maxIncrease = diff ;
}
if( diff > threshold){
increases++;
mem = newMem ;
}
else if(mem > newMem) {
mem = newMem ;
decreases++ ;
};
}
global.gc() ;
console.log(loopsToWatch + ":", increases + " increases",decreases + " decreases", " max " + maxIncrease," threshold " + threshold, " total " ,process.memoryUsage().heapUsed - intialMem," maxUsed " + maxUsed ) ;
```
Changing` loopsToWatch` , commenting out` global.gc` and/or switching the active line in` underTest()` all indicate an issue with `process.send`
with global.gc in loop...
```
>node parent
100: 96 increases 4 decreases max 216 threshold 32 total -392720 maxUsed 4292032
```
no global.gc in loop...
```
>node parent
100: 100 increases 0 decreases max 10056 threshold 32 total -94216 maxUsed 4741672
```
no global.gc in loop and 200 iterations...
```
>node parent
200: 200 increases 0 decreases max 10056 threshold 32 total -71032 maxUsed 4803616
```
looks like a limit reached at 101 reported increases with threshold at 1100.....
```
>node parent
200: 101 increases 0 decreases max 10056 threshold 1100 total -71000 maxUsed 4803592
```
but then...
```
>node parent
300: 151 increases 0 decreases max 10056 threshold 1100 total -51528 maxUsed 4865208
```
testing `readdirSync` instead
```
>node parent
200: 0 increases 3 decreases max 16 threshold 32 total -155712 maxUsed 4286992
```
Each call to process.send seems to eat up more and more heap it doesn't need (as it does the job on less here when forced to) until gc gets called which presumably doesn't normally happen in a synchronous loop.
|
2.0
|
process.send called in a loop in a forked child process eats up heap. - * **Version**: 8.9.1
* **Platform**: Windows 10 64 bit
* **Subsystem**: Child Process
<!-- Enter your issue details below this comment. -->
Similar-sounding issue [here](https://github.com/nodejs/node/issues/15651) but event based / not in a loop.
I've managed to reduce the used heap in a test by ~400K by forking with the` --expose-gc` option and calling `global.gc` after calling `process.send`.
parent.js ;
```
"use strict" ;
var forkOptions = {
execArgv:['--expose-gc']
};
const cp = require('child_process');
const child = cp.fork(`./child.js`,[],forkOptions);
child.on("message",function(m){
//console.log(m);
});
```
child.js ;
```
// invoke with --expose-gc
"use strict" ;
var fs = require('fs');
const loopsToWatch = 100 ;
const threshold = 32 ; // bytes
function underTest(){
// let newExtractDirFiles = fs.readdirSync("./"); // threshold = 32
process.send("yada");
}
let newMem = process.memoryUsage().heapUsed ;
let increases = 0 , decreases = 0 , maxIncrease = 0 , maxUsed = 0 ;
let mem = process.memoryUsage().heapUsed ;
let intialMem = mem ;
for(let i = 0 ; i < loopsToWatch ; i++){
underTest();
global.gc() ; //requires --expose-gc flag
newMem = process.memoryUsage().heapUsed ;
if(newMem > maxUsed){
maxUsed = newMem ;
}
let diff = newMem - mem ;
if(diff > maxIncrease){
maxIncrease = diff ;
}
if( diff > threshold){
increases++;
mem = newMem ;
}
else if(mem > newMem) {
mem = newMem ;
decreases++ ;
};
}
global.gc() ;
console.log(loopsToWatch + ":", increases + " increases",decreases + " decreases", " max " + maxIncrease," threshold " + threshold, " total " ,process.memoryUsage().heapUsed - intialMem," maxUsed " + maxUsed ) ;
```
Changing` loopsToWatch` , commenting out` global.gc` and/or switching the active line in` underTest()` all indicate an issue with `process.send`
with global.gc in loop...
```
>node parent
100: 96 increases 4 decreases max 216 threshold 32 total -392720 maxUsed 4292032
```
no global.gc in loop...
```
>node parent
100: 100 increases 0 decreases max 10056 threshold 32 total -94216 maxUsed 4741672
```
no global.gc in loop and 200 iterations...
```
>node parent
200: 200 increases 0 decreases max 10056 threshold 32 total -71032 maxUsed 4803616
```
looks like a limit reached at 101 reported increases with threshold at 1100.....
```
>node parent
200: 101 increases 0 decreases max 10056 threshold 1100 total -71000 maxUsed 4803592
```
but then...
```
>node parent
300: 151 increases 0 decreases max 10056 threshold 1100 total -51528 maxUsed 4865208
```
testing `readdirSync` instead
```
>node parent
200: 0 increases 3 decreases max 16 threshold 32 total -155712 maxUsed 4286992
```
Each call to process.send seems to eat up more and more heap it doesn't need (as it does the job on less here when forced to) until gc gets called which presumably doesn't normally happen in a synchronous loop.
|
process
|
process send called in a loop in a forked child process eats up heap version platform windows bit subsystem child process similar sounding issue but event based not in a loop i ve managed to reduce the used heap in a test by by forking with the expose gc option and calling global gc after calling process send parent js use strict var forkoptions execargv const cp require child process const child cp fork child js forkoptions child on message function m console log m child js invoke with expose gc use strict var fs require fs const loopstowatch const threshold bytes function undertest let newextractdirfiles fs readdirsync threshold process send yada let newmem process memoryusage heapused let increases decreases maxincrease maxused let mem process memoryusage heapused let intialmem mem for let i i loopstowatch i undertest global gc requires expose gc flag newmem process memoryusage heapused if newmem maxused maxused newmem let diff newmem mem if diff maxincrease maxincrease diff if diff threshold increases mem newmem else if mem newmem mem newmem decreases global gc console log loopstowatch increases increases decreases decreases max maxincrease threshold threshold total process memoryusage heapused intialmem maxused maxused changing loopstowatch commenting out global gc and or switching the active line in undertest all indicate an issue with process send with global gc in loop node parent increases decreases max threshold total maxused no global gc in loop node parent increases decreases max threshold total maxused no global gc in loop and iterations node parent increases decreases max threshold total maxused looks like a limit reached at reported increases with threshold at node parent increases decreases max threshold total maxused but then node parent increases decreases max threshold total maxused testing readdirsync instead node parent increases decreases max threshold total maxused each call to process send seems to eat up more and more heap it doesn t need as it does the job on less here when forced to until gc gets called which presumably doesn t normally happen in a synchronous loop
| 1
|
20,243
| 10,685,342,962
|
IssuesEvent
|
2019-10-22 12:30:07
|
citusdata/citus
|
https://api.github.com/repos/citusdata/citus
|
opened
|
Recursive planning should not pull unnecessary data in the target list
|
performance
|
In multi-shard SELECT queries, Citus is mostly smart enough to pull only the necessary data to the coordinator. However, with recursive planning, Citus
```SQL
-- have less verbose output
SET citus.shard_count TO 4;
CREATE TABLE users_table (user_id int, time timestamp, value_1 int, value_2 int, value_3 float, value_4 bigint);
SELECT create_distributed_table('users_table', 'user_id');
SET citus.log_remote_commands TO ON;
SET client_min_messages TO DEBUG;
-- query pushdown, we only need user_id in the coordinator, so it doesn't make sense
-- to pull other columns
-- as you can see on the queries sent to workers, we only do ` SELECT worker_column_1 AS user_id FROM `
-- meaning that the other column is not sent to the worker
-- Note: we use random to prevent postgresl to do the optimization (pull up the subquery)
SELECT user_id FROM (SELECT user_id, value_1 * random() FROM users_table) as foo;
LOG: issuing SELECT worker_column_1 AS user_id FROM (SELECT foo.user_id AS worker_column_1 FROM (SELECT users_table.user_id, ((users_table.value_1)::double precision OPERATOR(pg_catalog.*) random()) FROM sc1.users_table_102394 users_table) foo) worker_subquery
DETAIL: on server onderkalaci@localhost:9700
....
-- another query with recursive planning forced by OFFSET 0
-- this time the worker query includes both user id and
SELECT user_id FROM (SELECT user_id, value_1 * random() FROM users_table OFFSET 0) as foo;
LOG: issuing SELECT user_id, ((value_1)::double precision OPERATOR(pg_catalog.*) random()) FROM sc1.users_table_102394 users_table WHERE true
DETAIL: on server onderkalaci@localhost:9700
....
```
There are some basic prototypes can be found here: https://github.com/citusdata/citus/pull/2504/commits/00b93382940b4c27e5d0cedf176e5ed2de3596bf
|
True
|
Recursive planning should not pull unnecessary data in the target list - In multi-shard SELECT queries, Citus is mostly smart enough to pull only the necessary data to the coordinator. However, with recursive planning, Citus
```SQL
-- have less verbose output
SET citus.shard_count TO 4;
CREATE TABLE users_table (user_id int, time timestamp, value_1 int, value_2 int, value_3 float, value_4 bigint);
SELECT create_distributed_table('users_table', 'user_id');
SET citus.log_remote_commands TO ON;
SET client_min_messages TO DEBUG;
-- query pushdown, we only need user_id in the coordinator, so it doesn't make sense
-- to pull other columns
-- as you can see on the queries sent to workers, we only do ` SELECT worker_column_1 AS user_id FROM `
-- meaning that the other column is not sent to the worker
-- Note: we use random to prevent postgresl to do the optimization (pull up the subquery)
SELECT user_id FROM (SELECT user_id, value_1 * random() FROM users_table) as foo;
LOG: issuing SELECT worker_column_1 AS user_id FROM (SELECT foo.user_id AS worker_column_1 FROM (SELECT users_table.user_id, ((users_table.value_1)::double precision OPERATOR(pg_catalog.*) random()) FROM sc1.users_table_102394 users_table) foo) worker_subquery
DETAIL: on server onderkalaci@localhost:9700
....
-- another query with recursive planning forced by OFFSET 0
-- this time the worker query includes both user id and
SELECT user_id FROM (SELECT user_id, value_1 * random() FROM users_table OFFSET 0) as foo;
LOG: issuing SELECT user_id, ((value_1)::double precision OPERATOR(pg_catalog.*) random()) FROM sc1.users_table_102394 users_table WHERE true
DETAIL: on server onderkalaci@localhost:9700
....
```
There are some basic prototypes can be found here: https://github.com/citusdata/citus/pull/2504/commits/00b93382940b4c27e5d0cedf176e5ed2de3596bf
|
non_process
|
recursive planning should not pull unnecessary data in the target list in multi shard select queries citus is mostly smart enough to pull only the necessary data to the coordinator however with recursive planning citus sql have less verbose output set citus shard count to create table users table user id int time timestamp value int value int value float value bigint select create distributed table users table user id set citus log remote commands to on set client min messages to debug query pushdown we only need user id in the coordinator so it doesn t make sense to pull other columns as you can see on the queries sent to workers we only do select worker column as user id from meaning that the other column is not sent to the worker note we use random to prevent postgresl to do the optimization pull up the subquery select user id from select user id value random from users table as foo log issuing select worker column as user id from select foo user id as worker column from select users table user id users table value double precision operator pg catalog random from users table users table foo worker subquery detail on server onderkalaci localhost another query with recursive planning forced by offset this time the worker query includes both user id and select user id from select user id value random from users table offset as foo log issuing select user id value double precision operator pg catalog random from users table users table where true detail on server onderkalaci localhost there are some basic prototypes can be found here
| 0
|
11,081
| 13,921,248,581
|
IssuesEvent
|
2020-10-21 11:39:45
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Consolidate Undici error handling
|
process/candidate team/typescript
|
As of now, once in a while, here and there, an error [pops up](https://github.com/prisma/prisma/issues/3892)
We should go through all possible Undici errors and handle them properly.
|
1.0
|
Consolidate Undici error handling - As of now, once in a while, here and there, an error [pops up](https://github.com/prisma/prisma/issues/3892)
We should go through all possible Undici errors and handle them properly.
|
process
|
consolidate undici error handling as of now once in a while here and there an error we should go through all possible undici errors and handle them properly
| 1
|
683,210
| 23,372,119,236
|
IssuesEvent
|
2022-08-10 20:57:01
|
NeurodataWithoutBorders/pynwb
|
https://api.github.com/repos/NeurodataWithoutBorders/pynwb
|
closed
|
[Feature]: `ImageSeries`: check `starting_frame` when `external_file` is used
|
category: enhancement priority: high
|
### What would you like to see added to PyNWB?
Related to #1318 and #1470, when `external_file` is used in `ImageSeries` a new check should be added which ensures that
a) `starting_frame` is provided if `external_file` `is` used,
b) `starting_frame` has the same length as `external_file`.
Currently `starting_frame` is set to `[0]` by default, which is fine for a single external file but is incorrect for multiple external files.
### Is your feature request related to a problem?
_No response_
### What solution would you like?
A new check method in ImageSeries.__init__ that checks if
a) `starting_frame` is provided if external_file is used
b) `starting_frame` has the same length as external_file.
### Do you have any interest in helping implement the feature?
Yes.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/NeurodataWithoutBorders/pynwb/blob/dev/.github/CODE_OF_CONDUCT.rst)
- [X] Have you checked the [Contributing](https://github.com/NeurodataWithoutBorders/pynwb/blob/dev/docs/CONTRIBUTING.rst) document?
- [X] Have you ensured this change was not already [requested](https://github.com/NeurodataWithoutBorders/pynwb/issues)?
|
1.0
|
[Feature]: `ImageSeries`: check `starting_frame` when `external_file` is used - ### What would you like to see added to PyNWB?
Related to #1318 and #1470, when `external_file` is used in `ImageSeries` a new check should be added which ensures that
a) `starting_frame` is provided if `external_file` `is` used,
b) `starting_frame` has the same length as `external_file`.
Currently `starting_frame` is set to `[0]` by default, which is fine for a single external file but is incorrect for multiple external files.
### Is your feature request related to a problem?
_No response_
### What solution would you like?
A new check method in ImageSeries.__init__ that checks if
a) `starting_frame` is provided if external_file is used
b) `starting_frame` has the same length as external_file.
### Do you have any interest in helping implement the feature?
Yes.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/NeurodataWithoutBorders/pynwb/blob/dev/.github/CODE_OF_CONDUCT.rst)
- [X] Have you checked the [Contributing](https://github.com/NeurodataWithoutBorders/pynwb/blob/dev/docs/CONTRIBUTING.rst) document?
- [X] Have you ensured this change was not already [requested](https://github.com/NeurodataWithoutBorders/pynwb/issues)?
|
non_process
|
imageseries check starting frame when external file is used what would you like to see added to pynwb related to and when external file is used in imageseries a new check should be added which ensures that a starting frame is provided if external file is used b starting frame has the same length as external file currently starting frame is set to by default which is fine for a single external file but is incorrect for multiple external files is your feature request related to a problem no response what solution would you like a new check method in imageseries init that checks if a starting frame is provided if external file is used b starting frame has the same length as external file do you have any interest in helping implement the feature yes code of conduct i agree to follow this project s have you checked the document have you ensured this change was not already
| 0
|
9,467
| 12,464,219,994
|
IssuesEvent
|
2020-05-28 12:05:35
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
closed
|
DB::Exception: Block structure mismatch in QueryPipeline stream: different columns (Const)
|
bug comp-processors v20.3 v20.4
|
This simple view:
```
CREATE VIEW test.bug_const AS
SELECT 'World' AS hello
FROM (SELECT number FROM system.numbers LIMIT 1) AS n1
JOIN (SELECT number FROM system.numbers LIMIT 1) AS n2 USING (number)
```
is created ok, but at SELECT time throws the following exception:
> Received exception from server (version 20.4.3):
> Code: 171. DB::Exception: Received from localhost:9000. DB::Exception: Block structure mismatch in QueryPipeline stream: different columns:
> hello String String(size = 0)
> hello String Const(size = 0, String(size = 1)).
If I execute the SELECT part as as standalone query, it works, but in a VIEW it does not.
On complex views, joining other views many levels deep, this bug appears on all 20.3 and 20.4 versions of Clickhouse I tried, including 20.3.10.75, 20.4.2.9 and 20.4.3.16. But the simple way to reproduce it that I posted above only triggers the bug on 20.4 versions.
Branch 20.1 and previous ones seem to be OK.
|
1.0
|
DB::Exception: Block structure mismatch in QueryPipeline stream: different columns (Const) - This simple view:
```
CREATE VIEW test.bug_const AS
SELECT 'World' AS hello
FROM (SELECT number FROM system.numbers LIMIT 1) AS n1
JOIN (SELECT number FROM system.numbers LIMIT 1) AS n2 USING (number)
```
is created ok, but at SELECT time throws the following exception:
> Received exception from server (version 20.4.3):
> Code: 171. DB::Exception: Received from localhost:9000. DB::Exception: Block structure mismatch in QueryPipeline stream: different columns:
> hello String String(size = 0)
> hello String Const(size = 0, String(size = 1)).
If I execute the SELECT part as as standalone query, it works, but in a VIEW it does not.
On complex views, joining other views many levels deep, this bug appears on all 20.3 and 20.4 versions of Clickhouse I tried, including 20.3.10.75, 20.4.2.9 and 20.4.3.16. But the simple way to reproduce it that I posted above only triggers the bug on 20.4 versions.
Branch 20.1 and previous ones seem to be OK.
|
process
|
db exception block structure mismatch in querypipeline stream different columns const this simple view create view test bug const as select world as hello from select number from system numbers limit as join select number from system numbers limit as using number is created ok but at select time throws the following exception received exception from server version code db exception received from localhost db exception block structure mismatch in querypipeline stream different columns hello string string size hello string const size string size if i execute the select part as as standalone query it works but in a view it does not on complex views joining other views many levels deep this bug appears on all and versions of clickhouse i tried including and but the simple way to reproduce it that i posted above only triggers the bug on versions branch and previous ones seem to be ok
| 1
|
223,853
| 7,461,285,626
|
IssuesEvent
|
2018-03-31 00:56:33
|
containous/traefik
|
https://api.github.com/repos/containous/traefik
|
closed
|
Exempt the /metrics from HTTPS with Authentication
|
area/api area/middleware/metrics kind/enhancement priority/P3
|
Hi all,
I would like to request that if possible, exempt the /metrics from HTTPS with Basic Authentication if it's configured in [web] API.
The /metrics is for exporting internal metrics so that Prometheus can scrape on it. If the TLS certificate was a self-signed certificate, Prometheus will fail to scrape the /metrics from Traefik because it doesn't have an option to ignore for TLS certificate error (or maybe I'm not aware of its existence if it does exist).
Thank you in advance.
Kind regards,
Marvin
|
1.0
|
Exempt the /metrics from HTTPS with Authentication - Hi all,
I would like to request that if possible, exempt the /metrics from HTTPS with Basic Authentication if it's configured in [web] API.
The /metrics is for exporting internal metrics so that Prometheus can scrape on it. If the TLS certificate was a self-signed certificate, Prometheus will fail to scrape the /metrics from Traefik because it doesn't have an option to ignore for TLS certificate error (or maybe I'm not aware of its existence if it does exist).
Thank you in advance.
Kind regards,
Marvin
|
non_process
|
exempt the metrics from https with authentication hi all i would like to request that if possible exempt the metrics from https with basic authentication if it s configured in api the metrics is for exporting internal metrics so that prometheus can scrape on it if the tls certificate was a self signed certificate prometheus will fail to scrape the metrics from traefik because it doesn t have an option to ignore for tls certificate error or maybe i m not aware of its existence if it does exist thank you in advance kind regards marvin
| 0
|
22,059
| 30,575,854,434
|
IssuesEvent
|
2023-07-21 05:19:25
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Processors should remove SchemaURL when transformations violate semantic conventions
|
bug Stale processor/attributes processor/metricstransform processor/transform processor/cumulativetodelta processor/deltatorate processor/logstransform processor/resource closed as inactive
|
### Component(s)
processor/attributes, processor/cumulativetodelta, processor/deltatorate, processor/groupbyattrs, processor/logstransform, processor/metricstransform, processor/resource, processor/transform
### What happened?
## Description
When using a processor to alter telemetry that has a valid Telemetry Schema URL, it's possible/likely that the processor could cause the telemetry to violate the associated schema.
## Steps to Reproduce
- Create telemetry using https://github.com/open-telemetry/opentelemetry-specification/blob/main/schemas/1.20.0
- Manually rename attribute `net.protocol.name` back to `net.app.protocol.name` via the attributes processor
Telemetry now violates the schema
## Expected Result
The renamed span should be in a new InstrumentationScope that does NOT include a SchemaURL.
## Actual Result
The renamed span remains in an InstrumentationScope that declares compatibility with 1.20.0 of the semantic conventions.
### Collector version
v0.73.0
### Environment information
## Environment
The otel-collector-contrib docker image
### OpenTelemetry Collector configuration
```yaml
receivers:
otlp:
exporters:
logging:
processors:
attributes/bad:
actions:
- key: net.app.protocol.name
action: insert
from_context: net.protocol.name
- key: net.protocol.name
action: delete
pipelines:
traces:
receivers: [otlp]
processors: [attributes/bad]
exporters: [logging]
```
### Log output
_No response_
### Additional context
This is targeted at the viability of SchemaURL itself and its enforceability through the ecosystem.
- Can we simplify processing telemetry so SchemaURL is dropped whenever its guarantees would be violated?
- Do we have the right tools in place to determine, detect and enforce this?
- What does the collector need to make this portion of the specification as easy as possible to deal with?
cc @tigrannajaryan
|
7.0
|
Processors should remove SchemaURL when transformations violate semantic conventions - ### Component(s)
processor/attributes, processor/cumulativetodelta, processor/deltatorate, processor/groupbyattrs, processor/logstransform, processor/metricstransform, processor/resource, processor/transform
### What happened?
## Description
When using a processor to alter telemetry that has a valid Telemetry Schema URL, it's possible/likely that the processor could cause the telemetry to violate the associated schema.
## Steps to Reproduce
- Create telemetry using https://github.com/open-telemetry/opentelemetry-specification/blob/main/schemas/1.20.0
- Manually rename attribute `net.protocol.name` back to `net.app.protocol.name` via the attributes processor
Telemetry now violates the schema
## Expected Result
The renamed span should be in a new InstrumentationScope that does NOT include a SchemaURL.
## Actual Result
The renamed span remains in an InstrumentationScope that declares compatibility with 1.20.0 of the semantic conventions.
### Collector version
v0.73.0
### Environment information
## Environment
The otel-collector-contrib docker image
### OpenTelemetry Collector configuration
```yaml
receivers:
otlp:
exporters:
logging:
processors:
attributes/bad:
actions:
- key: net.app.protocol.name
action: insert
from_context: net.protocol.name
- key: net.protocol.name
action: delete
pipelines:
traces:
receivers: [otlp]
processors: [attributes/bad]
exporters: [logging]
```
### Log output
_No response_
### Additional context
This is targeted at the viability of SchemaURL itself and its enforceability through the ecosystem.
- Can we simplify processing telemetry so SchemaURL is dropped whenever its guarantees would be violated?
- Do we have the right tools in place to determine, detect and enforce this?
- What does the collector need to make this portion of the specification as easy as possible to deal with?
cc @tigrannajaryan
|
process
|
processors should remove schemaurl when transformations violate semantic conventions component s processor attributes processor cumulativetodelta processor deltatorate processor groupbyattrs processor logstransform processor metricstransform processor resource processor transform what happened description when using a processor to alter telemetry that has a valid telemetry schema url it s possible likely that the processor could cause the telemetry to violate the associated schema steps to reproduce create telemetry using manually rename attribute net protocol name back to net app protocol name via the attributes processor telemetry now violates the schema expected result the renamed span should be in a new instrumentationscope that does not include a schemaurl actual result the renamed span remains in an instrumentationscope that declares compatibility with of the semantic conventions collector version environment information environment the otel collector contrib docker image opentelemetry collector configuration yaml receivers otlp exporters logging processors attributes bad actions key net app protocol name action insert from context net protocol name key net protocol name action delete pipelines traces receivers processors exporters log output no response additional context this is targeted at the viability of schemaurl itself and its enforceability through the ecosystem can we simplify processing telemetry so schemaurl is dropped whenever its guarantees would be violated do we have the right tools in place to determine detect and enforce this what does the collector need to make this portion of the specification as easy as possible to deal with cc tigrannajaryan
| 1
|
14,713
| 17,926,559,793
|
IssuesEvent
|
2021-09-10 00:51:23
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Research about text recognition and ward detection scripts parallelization [2]
|
Data Processing Week 1 Research Sprint 4
|
Research if there is a way to use in parallel both of the scripts (text recognition and ward detection).
|
1.0
|
Research about text recognition and ward detection scripts parallelization [2] - Research if there is a way to use in parallel both of the scripts (text recognition and ward detection).
|
process
|
research about text recognition and ward detection scripts parallelization research if there is a way to use in parallel both of the scripts text recognition and ward detection
| 1
|
16,796
| 22,044,378,387
|
IssuesEvent
|
2022-05-29 21:02:11
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
The Adventure of Nathaniel and Isabel from Six Feet Under
|
suggested title in process
|
Please add as much of the following info as you can:
Title: The Adventure of Nathaniel and Isabel
Type (film/tv show): tv show
Film or show in which it appears: Six Feet Under
Is the parent film/show streaming anywhere? HBO MAX, Hulu.
About when in the parent film/show does it appear? Season 3 Episode 10 (~19:00 and ~28:00)
Actual footage of the film/show can be seen (yes/no)? yes
|
1.0
|
The Adventure of Nathaniel and Isabel from Six Feet Under - Please add as much of the following info as you can:
Title: The Adventure of Nathaniel and Isabel
Type (film/tv show): tv show
Film or show in which it appears: Six Feet Under
Is the parent film/show streaming anywhere? HBO MAX, Hulu.
About when in the parent film/show does it appear? Season 3 Episode 10 (~19:00 and ~28:00)
Actual footage of the film/show can be seen (yes/no)? yes
|
process
|
the adventure of nathaniel and isabel from six feet under please add as much of the following info as you can title the adventure of nathaniel and isabel type film tv show tv show film or show in which it appears six feet under is the parent film show streaming anywhere hbo max hulu about when in the parent film show does it appear season episode and actual footage of the film show can be seen yes no yes
| 1
|
18,116
| 24,148,578,422
|
IssuesEvent
|
2022-09-21 21:17:23
|
apache/arrow-datafusion
|
https://api.github.com/repos/apache/arrow-datafusion
|
closed
|
DataFusion 12.0.0 Release
|
enhancement development-process
|
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
Planning for DataFusion 12.0.0
- [x] https://github.com/apache/arrow-datafusion/issues/3192
- [x] https://github.com/apache/arrow-datafusion/pull/3225
- [x] https://github.com/apache/arrow-datafusion/pull/3407
- [x] https://github.com/apache/arrow-datafusion/pull/3398
**Describe the solution you'd like**
The usual schedule would have the RC cut on September 9th, but I plan on releasing it one week earlier, on September 2nd. This is ~partially due to the 9th being before the Labor Day weekend in the US, but also~ to line up with Dask SQL's planned GA release in September.
**Describe alternatives you've considered**
None
**Additional context**
None
|
1.0
|
DataFusion 12.0.0 Release - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
Planning for DataFusion 12.0.0
- [x] https://github.com/apache/arrow-datafusion/issues/3192
- [x] https://github.com/apache/arrow-datafusion/pull/3225
- [x] https://github.com/apache/arrow-datafusion/pull/3407
- [x] https://github.com/apache/arrow-datafusion/pull/3398
**Describe the solution you'd like**
The usual schedule would have the RC cut on September 9th, but I plan on releasing it one week earlier, on September 2nd. This is ~partially due to the 9th being before the Labor Day weekend in the US, but also~ to line up with Dask SQL's planned GA release in September.
**Describe alternatives you've considered**
None
**Additional context**
None
|
process
|
datafusion release is your feature request related to a problem or challenge please describe what you are trying to do planning for datafusion describe the solution you d like the usual schedule would have the rc cut on september but i plan on releasing it one week earlier on september this is partially due to the being before the labor day weekend in the us but also to line up with dask sql s planned ga release in september describe alternatives you ve considered none additional context none
| 1
|
34,683
| 14,485,757,817
|
IssuesEvent
|
2020-12-10 17:59:55
|
edgexfoundry/edgex-docs
|
https://api.github.com/repos/edgexfoundry/edgex-docs
|
closed
|
ADR for bootstrapping of EdgeX in OCI container environments
|
ADR security-services
|
## Problem statement
In Fuji release, we use Consul health checks to gate the initialization of security services. This works for the most part, but has a critical flaw: Consul itself is not secure. We need a bootstrapping mechanism that will allow us to secure all services, including Consul.
## Description
Write an ADR for bootstrapping of EdgeX in OCI container environments such as:
- Docker (docker-compose)
- Docker swarm
ADR should ensure that all EdgeX services can be brought up in a secure configuration:
- Third-party containers that can utilize TLS certificates will have the required TLS assets available at container start time
- Database accounts and passwords have been pre-seeded
- Paired services such as kong - postgres are set up in secure configuration
- Message busses are secured
ADR should define mutex/semaphore mechanisms to delay framework bringup until necessary initialization has completed.
|
1.0
|
ADR for bootstrapping of EdgeX in OCI container environments - ## Problem statement
In Fuji release, we use Consul health checks to gate the initialization of security services. This works for the most part, but has a critical flaw: Consul itself is not secure. We need a bootstrapping mechanism that will allow us to secure all services, including Consul.
## Description
Write an ADR for bootstrapping of EdgeX in OCI container environments such as:
- Docker (docker-compose)
- Docker swarm
ADR should ensure that all EdgeX services can be brought up in a secure configuration:
- Third-party containers that can utilize TLS certificates will have the required TLS assets available at container start time
- Database accounts and passwords have been pre-seeded
- Paired services such as kong - postgres are set up in secure configuration
- Message busses are secured
ADR should define mutex/semaphore mechanisms to delay framework bringup until necessary initialization has completed.
|
non_process
|
adr for bootstrapping of edgex in oci container environments problem statement in fuji release we use consul health checks to gate the initialization of security services this works for the most part but has a critical flaw consul itself is not secure we need a bootstrapping mechanism that will allow us to secure all services including consul description write an adr for bootstrapping of edgex in oci container environments such as docker docker compose docker swarm adr should ensure that all edgex services can be brought up in a secure configuration third party containers that can utilize tls certificates will have the required tls assets available at container start time database accounts and passwords have been pre seeded paired services such as kong postgres are set up in secure configuration message busses are secured adr should define mutex semaphore mechanisms to delay framework bringup until necessary initialization has completed
| 0
|
21,400
| 29,239,251,430
|
IssuesEvent
|
2023-05-23 00:18:09
|
ethereum/EIPs
|
https://api.github.com/repos/ethereum/EIPs
|
closed
|
"Requires" section should not accept `Withdrawn` EIP
|
w-stale enhancement r-ci r-process
|
### Proposed Change
The current EIPW bot allows `Final`, `Living` and `Withdrawn` EIP to be accepted in the "Requires" section.
While `Withdrawn` is a terminal status, it may not be an EIP which followed the standardization process and may not be a best proposal to be followed (and/or depend on).
Proposal: Active EIP should not be dependent on a `Withdrawn` EIP and not accepted in "Requires" section.
|
1.0
|
"Requires" section should not accept `Withdrawn` EIP - ### Proposed Change
The current EIPW bot allows `Final`, `Living` and `Withdrawn` EIP to be accepted in the "Requires" section.
While `Withdrawn` is a terminal status, it may not be an EIP which followed the standardization process and may not be a best proposal to be followed (and/or depend on).
Proposal: Active EIP should not be dependent on a `Withdrawn` EIP and not accepted in "Requires" section.
|
process
|
requires section should not accept withdrawn eip proposed change the current eipw bot allows final living and withdrawn eip to be accepted in the requires section while withdrawn is a terminal status it may not be an eip which followed the standardization process and may not be a best proposal to be followed and or depend on proposal active eip should not be dependent on a withdrawn eip and not accepted in requires section
| 1
|
2,325
| 5,142,121,262
|
IssuesEvent
|
2017-01-12 12:12:31
|
vuejs/vue-loader
|
https://api.github.com/repos/vuejs/vue-loader
|
reopened
|
Setting pug via webpack config won't render templates
|
pre-processor
|
I want to provide pug as template pre-processor in webpack config to avoid writing `lang="pug"` in every component.
My relevant chunk of Webpack config is:
```js
module: {
loaders: [{
test: /\.pug$/,
exclude: /node_modules/,
loader: 'pug',
}, {
test: /\.js$/,
exclude: /node_modules/,
loader: 'babel',
}, {
test: /\.vue$/,
exclude: /node_modules/,
loader: 'vue',
}],
},
vue: {
loaders: {
html: 'pug',
css: 'style!css!stylus',
},
},
```
I write components like this:
```html
<!-- index.vue -->
<template src="./template.pug"></template>
<script src="./script.js"></script>
<style src="./style.styl"></style>
```
The result is following error in console: `vue.js:2643[Vue warn]: Failed to mount component: template or render function not defined.`, and template won't render. However, Stylus style files are handled correctly without `lang="stylus"`. Is it bug or I may be missing something?
|
1.0
|
Setting pug via webpack config won't render templates - I want to provide pug as template pre-processor in webpack config to avoid writing `lang="pug"` in every component.
My relevant chunk of Webpack config is:
```js
module: {
loaders: [{
test: /\.pug$/,
exclude: /node_modules/,
loader: 'pug',
}, {
test: /\.js$/,
exclude: /node_modules/,
loader: 'babel',
}, {
test: /\.vue$/,
exclude: /node_modules/,
loader: 'vue',
}],
},
vue: {
loaders: {
html: 'pug',
css: 'style!css!stylus',
},
},
```
I write components like this:
```html
<!-- index.vue -->
<template src="./template.pug"></template>
<script src="./script.js"></script>
<style src="./style.styl"></style>
```
The result is following error in console: `vue.js:2643[Vue warn]: Failed to mount component: template or render function not defined.`, and template won't render. However, Stylus style files are handled correctly without `lang="stylus"`. Is it bug or I may be missing something?
|
process
|
setting pug via webpack config won t render templates i want to provide pug as template pre processor in webpack config to avoid writing lang pug in every component my relevant chunk of webpack config is js module loaders test pug exclude node modules loader pug test js exclude node modules loader babel test vue exclude node modules loader vue vue loaders html pug css style css stylus i write components like this html the result is following error in console vue js failed to mount component template or render function not defined and template won t render however stylus style files are handled correctly without lang stylus is it bug or i may be missing something
| 1
|
661,884
| 22,094,004,524
|
IssuesEvent
|
2022-06-01 08:30:46
|
PHI-base/PHI5_web_display
|
https://api.github.com/repos/PHI-base/PHI5_web_display
|
closed
|
The host gene pages should also show the metagenotypes and annotations
|
Priority
|


From this host gene page how can a user identify the corresponding pathogen avirulence effector eg 'AvrStb6'?
Currently one directional 'pathogen avr gene' -> 'host resistance gene'.
eg page https://poc.molecularconnections.com/phibase-v3/#/search-detail-page/PHIG:256
|
1.0
|
The host gene pages should also show the metagenotypes and annotations - 

From this host gene page how can a user identify the corresponding pathogen avirulence effector eg 'AvrStb6'?
Currently one directional 'pathogen avr gene' -> 'host resistance gene'.
eg page https://poc.molecularconnections.com/phibase-v3/#/search-detail-page/PHIG:256
|
non_process
|
the host gene pages should also show the metagenotypes and annotations from this host gene page how can a user identify the corresponding pathogen avirulence effector eg currently one directional pathogen avr gene host resistance gene eg page
| 0
|
120,768
| 4,793,832,805
|
IssuesEvent
|
2016-10-31 19:17:34
|
roschaefer/story.board
|
https://api.github.com/repos/roschaefer/story.board
|
opened
|
Color propertiy for triggers
|
Priority: medium
|
As a member of service team,
i want triggers to have a color-property,
to make them easier identifiable for reporters.
|
1.0
|
Color propertiy for triggers - As a member of service team,
i want triggers to have a color-property,
to make them easier identifiable for reporters.
|
non_process
|
color propertiy for triggers as a member of service team i want triggers to have a color property to make them easier identifiable for reporters
| 0
|
158,719
| 12,423,740,035
|
IssuesEvent
|
2020-05-24 07:37:33
|
ReactiveX/RxJava
|
https://api.github.com/repos/ReactiveX/RxJava
|
closed
|
3.x: Suppress undeliverables in some tests
|
3.x Cleanup PR welcome Test good first issue
|
Some tests trigger a lot of undeliverable exceptions flooding the console output. Example:
```
at io.reactivex.rxjava3.internal.subscribers.FutureSubscriberTest
.onErrorCancelRace(FutureSubscriberTest.java:178)
at io.reactivex.rxjava3.internal.operators.observable.ObservableWindowWithObservableTest
.disposeMainBoundaryErrorRace(ObservableWindowWithObservableTest.java:590)
```
These are generally irrelevant from those particular tests and should be safely ignored. The difficulty is, locating these tests if the console doesn't retain too many lines.
To resolve the issue, perhaps the easiest way is to write a `TestHelper.withUndeliverableSuppressed(Action)` method and keep wrapping the problematic test bodies.
|
1.0
|
3.x: Suppress undeliverables in some tests - Some tests trigger a lot of undeliverable exceptions flooding the console output. Example:
```
at io.reactivex.rxjava3.internal.subscribers.FutureSubscriberTest
.onErrorCancelRace(FutureSubscriberTest.java:178)
at io.reactivex.rxjava3.internal.operators.observable.ObservableWindowWithObservableTest
.disposeMainBoundaryErrorRace(ObservableWindowWithObservableTest.java:590)
```
These are generally irrelevant from those particular tests and should be safely ignored. The difficulty is, locating these tests if the console doesn't retain too many lines.
To resolve the issue, perhaps the easiest way is to write a `TestHelper.withUndeliverableSuppressed(Action)` method and keep wrapping the problematic test bodies.
|
non_process
|
x suppress undeliverables in some tests some tests trigger a lot of undeliverable exceptions flooding the console output example at io reactivex internal subscribers futuresubscribertest onerrorcancelrace futuresubscribertest java at io reactivex internal operators observable observablewindowwithobservabletest disposemainboundaryerrorrace observablewindowwithobservabletest java these are generally irrelevant from those particular tests and should be safely ignored the difficulty is locating these tests if the console doesn t retain too many lines to resolve the issue perhaps the easiest way is to write a testhelper withundeliverablesuppressed action method and keep wrapping the problematic test bodies
| 0
|
323,080
| 9,842,712,742
|
IssuesEvent
|
2019-06-18 09:52:37
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
opened
|
Changing skaffold.yaml stops `skaffold dev`
|
kind/bug priority/p0
|
Changing the content of `skaffold.yaml` used to restart the dev loop. Not, it exits with this error:
```
FATA[0197] calling final callback: configuration changed
```
|
1.0
|
Changing skaffold.yaml stops `skaffold dev` - Changing the content of `skaffold.yaml` used to restart the dev loop. Not, it exits with this error:
```
FATA[0197] calling final callback: configuration changed
```
|
non_process
|
changing skaffold yaml stops skaffold dev changing the content of skaffold yaml used to restart the dev loop not it exits with this error fata calling final callback configuration changed
| 0
|
18,540
| 24,554,756,465
|
IssuesEvent
|
2022-10-12 15:03:48
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Android] Activities scheduling is not matching with the SB configuration
|
Bug Blocker P0 Android Process: Fixed Process: Tested QA Process: Tested dev
|
Steps:-
1. Configure the multiple activities with different scheduling types and Options and Publish
2. Install and Login into the mobile app
3. Enroll into the study and navigate to Study activities screen
4. Compare the schedule(runs) available for each activity
**A/R:-** Currently, mismatch among the activity and runs. **Eg:-** Runs for activity-1 is displaying for activity-2 and activity status is not updating as expected
**E/R:-** Scheduling should work as configured in SB and activity status should be updated properly
Studyname - **FHIR_DID-Study (All response types)[copied0208]**
Study ID - **copied0208**

|
3.0
|
[Android] Activities scheduling is not matching with the SB configuration - Steps:-
1. Configure the multiple activities with different scheduling types and Options and Publish
2. Install and Login into the mobile app
3. Enroll into the study and navigate to Study activities screen
4. Compare the schedule(runs) available for each activity
**A/R:-** Currently, mismatch among the activity and runs. **Eg:-** Runs for activity-1 is displaying for activity-2 and activity status is not updating as expected
**E/R:-** Scheduling should work as configured in SB and activity status should be updated properly
Studyname - **FHIR_DID-Study (All response types)[copied0208]**
Study ID - **copied0208**

|
process
|
activities scheduling is not matching with the sb configuration steps configure the multiple activities with different scheduling types and options and publish install and login into the mobile app enroll into the study and navigate to study activities screen compare the schedule runs available for each activity a r currently mismatch among the activity and runs eg runs for activity is displaying for activity and activity status is not updating as expected e r scheduling should work as configured in sb and activity status should be updated properly studyname fhir did study all response types study id
| 1
|
13,819
| 16,581,964,330
|
IssuesEvent
|
2021-05-31 13:04:45
|
jessestewart1/nrn-rrn
|
https://api.github.com/repos/jessestewart1/nrn-rrn
|
opened
|
Process PE 2021 - reissue with new addresses
|
processing
|
**Description of tasks**
Process PE 2021 data for release as an NRN product. This comes from a request by the data provider who wishes to see some recent address changes reflected on the NRN as the original addresses provided were outdated.
- [ ] update field mapping yaml(s)
- [ ] process PE 2021 data
- [ ] update release notes and sphinx documentation
- [ ] copy data to server
- [ ] confirm WMS updates and publication to Open Maps
- [ ] custom task: merge output with source data attributes.
|
1.0
|
Process PE 2021 - reissue with new addresses - **Description of tasks**
Process PE 2021 data for release as an NRN product. This comes from a request by the data provider who wishes to see some recent address changes reflected on the NRN as the original addresses provided were outdated.
- [ ] update field mapping yaml(s)
- [ ] process PE 2021 data
- [ ] update release notes and sphinx documentation
- [ ] copy data to server
- [ ] confirm WMS updates and publication to Open Maps
- [ ] custom task: merge output with source data attributes.
|
process
|
process pe reissue with new addresses description of tasks process pe data for release as an nrn product this comes from a request by the data provider who wishes to see some recent address changes reflected on the nrn as the original addresses provided were outdated update field mapping yaml s process pe data update release notes and sphinx documentation copy data to server confirm wms updates and publication to open maps custom task merge output with source data attributes
| 1
|
16,149
| 20,508,290,248
|
IssuesEvent
|
2022-03-01 01:47:48
|
g4he/g4he
|
https://api.github.com/repos/g4he/g4he
|
closed
|
Build a new index
|
feature: data processing
|
Now that the indexer has been updated, we should build a new index - do not overwrite the current one or any older ones, keep them for later analysis. Make a new index with a new name and update the frontend service to read from the new index endpoint.
Note as per issue #71 that once it is rebuilt we should check the reports still work properly, and fix any references to old keys.
|
1.0
|
Build a new index - Now that the indexer has been updated, we should build a new index - do not overwrite the current one or any older ones, keep them for later analysis. Make a new index with a new name and update the frontend service to read from the new index endpoint.
Note as per issue #71 that once it is rebuilt we should check the reports still work properly, and fix any references to old keys.
|
process
|
build a new index now that the indexer has been updated we should build a new index do not overwrite the current one or any older ones keep them for later analysis make a new index with a new name and update the frontend service to read from the new index endpoint note as per issue that once it is rebuilt we should check the reports still work properly and fix any references to old keys
| 1
|
489,892
| 14,113,286,433
|
IssuesEvent
|
2020-11-07 10:24:51
|
WoWManiaUK/Redemption
|
https://api.github.com/repos/WoWManiaUK/Redemption
|
closed
|
[Dungeon] Pit of Saron entrance event missing
|
Confirmed - Improvement needed Priority - Low
|
**Links:** https://youtu.be/C2BOcAbvVkA?t=23 to 0:50
**What is Happening:** The roleplay/event upon entering Pit of Saron where the Coliseum Champions of the Horde/Alliance get killed by Tyrannus and converted into skeletons before aggroing the party is missing. The NPC formation is present right now, but no event plays.
**What Should happen:** See video above from 0:23 to 0:50. Allied NPCs should get sucked and converted to undead, before engaging the party.
|
1.0
|
[Dungeon] Pit of Saron entrance event missing - **Links:** https://youtu.be/C2BOcAbvVkA?t=23 to 0:50
**What is Happening:** The roleplay/event upon entering Pit of Saron where the Coliseum Champions of the Horde/Alliance get killed by Tyrannus and converted into skeletons before aggroing the party is missing. The NPC formation is present right now, but no event plays.
**What Should happen:** See video above from 0:23 to 0:50. Allied NPCs should get sucked and converted to undead, before engaging the party.
|
non_process
|
pit of saron entrance event missing links to what is happening the roleplay event upon entering pit of saron where the coliseum champions of the horde alliance get killed by tyrannus and converted into skeletons before aggroing the party is missing the npc formation is present right now but no event plays what should happen see video above from to allied npcs should get sucked and converted to undead before engaging the party
| 0
|
2,384
| 5,187,641,230
|
IssuesEvent
|
2017-01-20 17:24:40
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
closed
|
Tabs not displayed in completed start form
|
browser: all bug comp: activiti-processList
|
1. Start a process that has a form attached to its start event, ensure form has more than one tab
2. Complete start event
3. Complete process
4. Go to 'completed' filter
5. Go to completed start form
**Expected results**
All tabs should be displayed
**Actual results**
Only first tab is displayed correctly
<img width="627" alt="screen shot 2017-01-09 at 11 24 25" src="https://cloud.githubusercontent.com/assets/13200338/21765118/b9ce83e0-d65e-11e6-9d37-115ef4d5679c.png">
|
1.0
|
Tabs not displayed in completed start form - 1. Start a process that has a form attached to its start event, ensure form has more than one tab
2. Complete start event
3. Complete process
4. Go to 'completed' filter
5. Go to completed start form
**Expected results**
All tabs should be displayed
**Actual results**
Only first tab is displayed correctly
<img width="627" alt="screen shot 2017-01-09 at 11 24 25" src="https://cloud.githubusercontent.com/assets/13200338/21765118/b9ce83e0-d65e-11e6-9d37-115ef4d5679c.png">
|
process
|
tabs not displayed in completed start form start a process that has a form attached to its start event ensure form has more than one tab complete start event complete process go to completed filter go to completed start form expected results all tabs should be displayed actual results only first tab is displayed correctly img width alt screen shot at src
| 1
|
17,079
| 22,582,222,777
|
IssuesEvent
|
2022-06-28 12:39:24
|
alphagov/govuk-design-system
|
https://api.github.com/repos/alphagov/govuk-design-system
|
opened
|
Design and purchase swag for Design System Day 2022
|
🕔 weeks process shared ownership
|
## What
Purchase swag for attendees of Design System Day 2022, now we have spending approval.
Continues from #2167.
## Why
So we can give attendees something following the event.
## Who needs to work on this
Designers
## Done when
- [ ] Decided on what items we're including
- [ ] Designs produced for each item (where applicable)
- [ ] Items have been ordered
- [ ] Items have been received for distribution
|
1.0
|
Design and purchase swag for Design System Day 2022 - ## What
Purchase swag for attendees of Design System Day 2022, now we have spending approval.
Continues from #2167.
## Why
So we can give attendees something following the event.
## Who needs to work on this
Designers
## Done when
- [ ] Decided on what items we're including
- [ ] Designs produced for each item (where applicable)
- [ ] Items have been ordered
- [ ] Items have been received for distribution
|
process
|
design and purchase swag for design system day what purchase swag for attendees of design system day now we have spending approval continues from why so we can give attendees something following the event who needs to work on this designers done when decided on what items we re including designs produced for each item where applicable items have been ordered items have been received for distribution
| 1
|
264,663
| 20,027,625,497
|
IssuesEvent
|
2022-02-01 23:30:59
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
[Docs] Plugin discovery system stops scanning when a plugin manifest is found
|
Feature:Plugins Team:Core docs DevDocs documentation
|
In https://github.com/elastic/kibana/blob/main/dev_docs/key_concepts/anatomy_of_a_plugin.mdx we cover the basic anatomy of a plugin and explain how plugins interact with each other and core.
A number of folks have asked about nesting plugins inside other plugins and how that works. We need to add a section that explains the concept of "sub-plugins":
One can have a folder with multiple plugins with the restriction that one can't have plugins nested inside other plugins because the plugin discovery system stops scanning recursively when a plugin manifest is found.
|
1.0
|
[Docs] Plugin discovery system stops scanning when a plugin manifest is found - In https://github.com/elastic/kibana/blob/main/dev_docs/key_concepts/anatomy_of_a_plugin.mdx we cover the basic anatomy of a plugin and explain how plugins interact with each other and core.
A number of folks have asked about nesting plugins inside other plugins and how that works. We need to add a section that explains the concept of "sub-plugins":
One can have a folder with multiple plugins with the restriction that one can't have plugins nested inside other plugins because the plugin discovery system stops scanning recursively when a plugin manifest is found.
|
non_process
|
plugin discovery system stops scanning when a plugin manifest is found in we cover the basic anatomy of a plugin and explain how plugins interact with each other and core a number of folks have asked about nesting plugins inside other plugins and how that works we need to add a section that explains the concept of sub plugins one can have a folder with multiple plugins with the restriction that one can t have plugins nested inside other plugins because the plugin discovery system stops scanning recursively when a plugin manifest is found
| 0
|
447,842
| 12,906,000,269
|
IssuesEvent
|
2020-07-15 00:11:21
|
knative/docs
|
https://api.github.com/repos/knative/docs
|
reopened
|
Document which PodSpec fields are valid
|
kind/serving lifecycle/rotten priority/1
|
**Describe the change you'd like to see**
While the yaml for a KnService leverages the Kubernetes PodSpec definition, not all of the PodSpec fields are valid for KnServices. We should document which fields people can use.
|
1.0
|
Document which PodSpec fields are valid - **Describe the change you'd like to see**
While the yaml for a KnService leverages the Kubernetes PodSpec definition, not all of the PodSpec fields are valid for KnServices. We should document which fields people can use.
|
non_process
|
document which podspec fields are valid describe the change you d like to see while the yaml for a knservice leverages the kubernetes podspec definition not all of the podspec fields are valid for knservices we should document which fields people can use
| 0
|
35,689
| 9,640,121,482
|
IssuesEvent
|
2019-05-16 14:53:02
|
fritzing/fritzing-app
|
https://api.github.com/repos/fritzing/fritzing-app
|
closed
|
Make sure to use release version of libgit2
|
CI / Delivery / Build system
|
In the develop and fix-windows-ci branches, we automatically build the libgit2 on travis and appveyor.
Unexpectedly, git2.dll ends up in build64/Debug folder. Thus, we build a debug version, which we don't need, plus, our scripts are confused about where to find the dll.
Solution might be to add `--config release`
See also
https://stackoverflow.com/questions/49417434/libgit2-is-not-building-in-release-mode-on-windows
|
1.0
|
Make sure to use release version of libgit2 - In the develop and fix-windows-ci branches, we automatically build the libgit2 on travis and appveyor.
Unexpectedly, git2.dll ends up in build64/Debug folder. Thus, we build a debug version, which we don't need, plus, our scripts are confused about where to find the dll.
Solution might be to add `--config release`
See also
https://stackoverflow.com/questions/49417434/libgit2-is-not-building-in-release-mode-on-windows
|
non_process
|
make sure to use release version of in the develop and fix windows ci branches we automatically build the on travis and appveyor unexpectedly dll ends up in debug folder thus we build a debug version which we don t need plus our scripts are confused about where to find the dll solution might be to add config release see also
| 0
|
17,784
| 23,713,755,632
|
IssuesEvent
|
2022-08-30 09:59:02
|
bjorkgard/public-secretary
|
https://api.github.com/repos/bjorkgard/public-secretary
|
closed
|
Exportera listor med förordnade förkunnare
|
Publisher in process
|
Användare skall ha rättigheter
Välja vilken typ av förordnande som skall kunna exporteras
- Äldste
- Församlingstjänare
- Äldste & Församlingstjänare
- Pionjärer & Specialpionjärer
|
1.0
|
Exportera listor med förordnade förkunnare - Användare skall ha rättigheter
Välja vilken typ av förordnande som skall kunna exporteras
- Äldste
- Församlingstjänare
- Äldste & Församlingstjänare
- Pionjärer & Specialpionjärer
|
process
|
exportera listor med förordnade förkunnare användare skall ha rättigheter välja vilken typ av förordnande som skall kunna exporteras äldste församlingstjänare äldste församlingstjänare pionjärer specialpionjärer
| 1
|
63,034
| 3,193,881,190
|
IssuesEvent
|
2015-09-30 08:49:44
|
fusioninventory/fusioninventory-for-glpi
|
https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi
|
closed
|
Not use PLUGIN_FUSIONINVENTORY_VERSION in fusinvdeploy plugin
|
Component: For junior contributor Component: Found in version Priority: Normal Status: Closed Tracker: Bug
|
---
Author Name: **David Durieux** (@ddurieux)
Original Redmine Issue: 1491, http://forge.fusioninventory.org/issues/1491
Original Date: 2012-02-23
Original Assignee: Gonéri Le Bouder
---
Not use this variable because on some case not initialized when try to use it in this plugin.
(error detected by unit tests)
|
1.0
|
Not use PLUGIN_FUSIONINVENTORY_VERSION in fusinvdeploy plugin - ---
Author Name: **David Durieux** (@ddurieux)
Original Redmine Issue: 1491, http://forge.fusioninventory.org/issues/1491
Original Date: 2012-02-23
Original Assignee: Gonéri Le Bouder
---
Not use this variable because on some case not initialized when try to use it in this plugin.
(error detected by unit tests)
|
non_process
|
not use plugin fusioninventory version in fusinvdeploy plugin author name david durieux ddurieux original redmine issue original date original assignee gonéri le bouder not use this variable because on some case not initialized when try to use it in this plugin error detected by unit tests
| 0
|
124,430
| 10,311,298,466
|
IssuesEvent
|
2019-08-29 17:01:16
|
aragon/aragon-court
|
https://api.github.com/repos/aragon/aragon-court
|
closed
|
Add tests for jurors registry drafting
|
component:tests
|
We need to add tests for `draft` and `slash` functions of the `JurorsRegistry`
|
1.0
|
Add tests for jurors registry drafting - We need to add tests for `draft` and `slash` functions of the `JurorsRegistry`
|
non_process
|
add tests for jurors registry drafting we need to add tests for draft and slash functions of the jurorsregistry
| 0
|
20,876
| 27,662,509,914
|
IssuesEvent
|
2023-03-12 17:31:33
|
LLazyEmail/nomoretogo_email_template
|
https://api.github.com/repos/LLazyEmail/nomoretogo_email_template
|
closed
|
running scripts gets confusing
|
in process
|
when i need to add or test some stuff, i'm adding another line in package json while debugging.
now, i'm putting things into bash folder

|
1.0
|
running scripts gets confusing - when i need to add or test some stuff, i'm adding another line in package json while debugging.
now, i'm putting things into bash folder

|
process
|
running scripts gets confusing when i need to add or test some stuff i m adding another line in package json while debugging now i m putting things into bash folder
| 1
|
9,226
| 12,259,150,021
|
IssuesEvent
|
2020-05-06 16:08:50
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Mediated trade shows locked funds warning
|
an:investigation in:arbitration in:trade-process was:dropped
|
Hi,
I have a mediated trade from a few weeks ago (17th Oct 2019), in my logs its saying i have locked funds for this trade still?
```
bisq.core.trade.TradeManager: We found a closed trade with locked up funds. That should never happen. trade ID=czvnyxa-XXX
```
Mediator suggested a payout which was accepted.
```json
"mediatorNodeAddress": {
"hostName": "exihmfza5343eh7q.onion",
"port": 9999
},
```
Let me know if you need more info.
This is potentially just a local wallet state issue since i sometimes send from my Bisq wallet with electrum.
|
1.0
|
Mediated trade shows locked funds warning - Hi,
I have a mediated trade from a few weeks ago (17th Oct 2019), in my logs its saying i have locked funds for this trade still?
```
bisq.core.trade.TradeManager: We found a closed trade with locked up funds. That should never happen. trade ID=czvnyxa-XXX
```
Mediator suggested a payout which was accepted.
```json
"mediatorNodeAddress": {
"hostName": "exihmfza5343eh7q.onion",
"port": 9999
},
```
Let me know if you need more info.
This is potentially just a local wallet state issue since i sometimes send from my Bisq wallet with electrum.
|
process
|
mediated trade shows locked funds warning hi i have a mediated trade from a few weeks ago oct in my logs its saying i have locked funds for this trade still bisq core trade trademanager we found a closed trade with locked up funds that should never happen trade id czvnyxa xxx mediator suggested a payout which was accepted json mediatornodeaddress hostname onion port let me know if you need more info this is potentially just a local wallet state issue since i sometimes send from my bisq wallet with electrum
| 1
|
77,881
| 15,569,901,434
|
IssuesEvent
|
2021-03-17 01:15:29
|
ScottCampit/stats503-w20-group17-project
|
https://api.github.com/repos/ScottCampit/stats503-w20-group17-project
|
opened
|
CVE-2020-11538 (High) detected in Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2020-11538 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: stats503-w20-group17-project/webapp/requirements.txt</p>
<p>Path to vulnerable library: stats503-w20-group17-project/webapp/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In libImaging/SgiRleDecode.c in Pillow through 7.0.0, a number of out-of-bounds reads exist in the parsing of SGI image files, a different issue than CVE-2020-5311.
<p>Publish Date: 2020-06-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11538>CVE-2020-11538</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p>
<p>Release Date: 2020-06-25</p>
<p>Fix Resolution: 7.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11538 (High) detected in Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2020-11538 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: stats503-w20-group17-project/webapp/requirements.txt</p>
<p>Path to vulnerable library: stats503-w20-group17-project/webapp/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In libImaging/SgiRleDecode.c in Pillow through 7.0.0, a number of out-of-bounds reads exist in the parsing of SGI image files, a different issue than CVE-2020-5311.
<p>Publish Date: 2020-06-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11538>CVE-2020-11538</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p>
<p>Release Date: 2020-06-25</p>
<p>Fix Resolution: 7.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file project webapp requirements txt path to vulnerable library project webapp requirements txt dependency hierarchy x pillow whl vulnerable library vulnerability details in libimaging sgirledecode c in pillow through a number of out of bounds reads exist in the parsing of sgi image files a different issue than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
21,443
| 29,478,670,146
|
IssuesEvent
|
2023-06-02 02:10:43
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Plugin is broken after upgrade to TS 4.0 (or Angular 10)
|
npm: @cypress/webpack-preprocessor stale
|
<!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. -->
- Operating System: macos
- Cypress Version:"@nrwl/cypress": "10.2.1",
- Browser Version: Updating Google Chrome
Version 85.0.4183.102 (Official Build) (64-bit)
### Is this a Feature or Bug?
Feature
### Current behavior:
Withing TS 3 and Angular 9 cypress works fine
Whenever upgrade one of two, we get broken. Probably due to TS 4. We could try to solve it out by adding peerDependency as a hotfix to package.json
### Desired behavior:
lib goes hand-in-hand with community
### How to reproduce:
use ts 4
### Additional Info (images, stack traces, etc)
|
1.0
|
Plugin is broken after upgrade to TS 4.0 (or Angular 10) - <!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. -->
- Operating System: macos
- Cypress Version:"@nrwl/cypress": "10.2.1",
- Browser Version: Updating Google Chrome
Version 85.0.4183.102 (Official Build) (64-bit)
### Is this a Feature or Bug?
Feature
### Current behavior:
Withing TS 3 and Angular 9 cypress works fine
Whenever upgrade one of two, we get broken. Probably due to TS 4. We could try to solve it out by adding peerDependency as a hotfix to package.json
### Desired behavior:
lib goes hand-in-hand with community
### How to reproduce:
use ts 4
### Additional Info (images, stack traces, etc)
|
process
|
plugin is broken after upgrade to ts or angular operating system macos cypress version nrwl cypress browser version updating google chrome version official build bit is this a feature or bug feature current behavior withing ts and angular cypress works fine whenever upgrade one of two we get broken probably due to ts we could try to solve it out by adding peerdependency as a hotfix to package json desired behavior lib goes hand in hand with community how to reproduce use ts additional info images stack traces etc
| 1
|
56,411
| 6,978,662,443
|
IssuesEvent
|
2017-12-12 18:16:49
|
SharePoint/sp-dev-docs
|
https://api.github.com/repos/SharePoint/sp-dev-docs
|
closed
|
SPFx 1.4 web part does no appear in the web part gallery
|
status:by-design
|
#### Category
- [ ] Question
- [ ] Typo
- [x] Bug
- [ ] Additional article idea
#### Expected or Desired Behavior
Being able to add the web part to the page after from the app catalog
#### Observed Behavior
Updated the framework to version 1.4 to test the asset packaging and after it I can't see the web parts that I install in the web part gallery.
Tested with the default hello world, it runs fine on both workbenches but after being installed it does not appear in the web part gallery.
Used the global app catalog and the site collection app catalog and got the same behaviour, the app is deployed with **No Errors**
Tested the **skipFeatureDeployment** on and off and the got the same behaviour, even after adding the app manually on site contents.
#### Steps to Reproduce
1. Create the hello-world default SPFx web part
2. Make sure the **includeClientSideAssets** is set to true
3. Run the command gulp bundle --ship
4. Run the command gulp package-solution --ship
5. Add the package to the app catalog (local or site collection)
6. Go to a modern page and check if your web part is available
|
1.0
|
SPFx 1.4 web part does no appear in the web part gallery - #### Category
- [ ] Question
- [ ] Typo
- [x] Bug
- [ ] Additional article idea
#### Expected or Desired Behavior
Being able to add the web part to the page after from the app catalog
#### Observed Behavior
Updated the framework to version 1.4 to test the asset packaging and after it I can't see the web parts that I install in the web part gallery.
Tested with the default hello world, it runs fine on both workbenches but after being installed it does not appear in the web part gallery.
Used the global app catalog and the site collection app catalog and got the same behaviour, the app is deployed with **No Errors**
Tested the **skipFeatureDeployment** on and off and the got the same behaviour, even after adding the app manually on site contents.
#### Steps to Reproduce
1. Create the hello-world default SPFx web part
2. Make sure the **includeClientSideAssets** is set to true
3. Run the command gulp bundle --ship
4. Run the command gulp package-solution --ship
5. Add the package to the app catalog (local or site collection)
6. Go to a modern page and check if your web part is available
|
non_process
|
spfx web part does no appear in the web part gallery category question typo bug additional article idea expected or desired behavior being able to add the web part to the page after from the app catalog observed behavior updated the framework to version to test the asset packaging and after it i can t see the web parts that i install in the web part gallery tested with the default hello world it runs fine on both workbenches but after being installed it does not appear in the web part gallery used the global app catalog and the site collection app catalog and got the same behaviour the app is deployed with no errors tested the skipfeaturedeployment on and off and the got the same behaviour even after adding the app manually on site contents steps to reproduce create the hello world default spfx web part make sure the includeclientsideassets is set to true run the command gulp bundle ship run the command gulp package solution ship add the package to the app catalog local or site collection go to a modern page and check if your web part is available
| 0
|
22,184
| 30,733,353,735
|
IssuesEvent
|
2023-07-28 05:04:28
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
Multiprocessing not propagating -Xfrozen_modules=off
|
type-bug 3.11 3.12 topic-multiprocessing
|
# Bug report
After python3.11 changes around frozen imports, when using multiprocessing contexts other than `fork`, the newly added `-Xfrozen_modules=off` isn't passed to spawned process interpreters.
Simple snippet demonstrating the issue:
```python
"""
$ python -Xfrozen_modules=off test.py
main: {'frozen_modules': 'off'}
forkserver: {}
spawn: {}
fork: {'frozen_modules': 'off'}
"""
import sys
import multiprocessing
def xoptions():
return sys._xoptions
def main():
print('main:', xoptions())
for ctx in ('forkserver', 'spawn', 'fork'):
with multiprocessing.get_context(ctx).Pool(1) as pool:
print(f'{ctx}:', pool.apply(xoptions))
if __name__ == '__main__':
main()
```
The issue seems to be `subprocess._args_from_interpreter_flags` not honoring `frozen_modules` key from `sys._xoptions`.
```bash
$ python -Xfrozen_modules=off -c 'import subprocess;print(subprocess._args_from_interpreter_flags())'
[]
```
# Your environment
python 3.11.4
<!-- gh-linked-prs -->
### Linked PRs
* gh-106724
<!-- /gh-linked-prs -->
|
1.0
|
Multiprocessing not propagating -Xfrozen_modules=off - # Bug report
After python3.11 changes around frozen imports, when using multiprocessing contexts other than `fork`, the newly added `-Xfrozen_modules=off` isn't passed to spawned process interpreters.
Simple snippet demonstrating the issue:
```python
"""
$ python -Xfrozen_modules=off test.py
main: {'frozen_modules': 'off'}
forkserver: {}
spawn: {}
fork: {'frozen_modules': 'off'}
"""
import sys
import multiprocessing
def xoptions():
return sys._xoptions
def main():
print('main:', xoptions())
for ctx in ('forkserver', 'spawn', 'fork'):
with multiprocessing.get_context(ctx).Pool(1) as pool:
print(f'{ctx}:', pool.apply(xoptions))
if __name__ == '__main__':
main()
```
The issue seems to be `subprocess._args_from_interpreter_flags` not honoring `frozen_modules` key from `sys._xoptions`.
```bash
$ python -Xfrozen_modules=off -c 'import subprocess;print(subprocess._args_from_interpreter_flags())'
[]
```
# Your environment
python 3.11.4
<!-- gh-linked-prs -->
### Linked PRs
* gh-106724
<!-- /gh-linked-prs -->
|
process
|
multiprocessing not propagating xfrozen modules off bug report after changes around frozen imports when using multiprocessing contexts other than fork the newly added xfrozen modules off isn t passed to spawned process interpreters simple snippet demonstrating the issue python python xfrozen modules off test py main frozen modules off forkserver spawn fork frozen modules off import sys import multiprocessing def xoptions return sys xoptions def main print main xoptions for ctx in forkserver spawn fork with multiprocessing get context ctx pool as pool print f ctx pool apply xoptions if name main main the issue seems to be subprocess args from interpreter flags not honoring frozen modules key from sys xoptions bash python xfrozen modules off c import subprocess print subprocess args from interpreter flags your environment python linked prs gh
| 1
|
18,124
| 24,163,387,183
|
IssuesEvent
|
2022-09-22 13:23:09
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Obsolete GO:0060883 regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication
|
obsoletion ready cellular processes
|
Dear all,
The proposal has been made to obsolete GO:0060883 regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication. The reason for obsoletion is that this represents a precomposed term. There are no annotations or mappings to this term; this term is not present in any subsets.
You can comment on the ticket:
Thanks, Pascale
|
1.0
|
Obsolete GO:0060883 regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication - Dear all,
The proposal has been made to obsolete GO:0060883 regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication. The reason for obsoletion is that this represents a precomposed term. There are no annotations or mappings to this term; this term is not present in any subsets.
You can comment on the ticket:
Thanks, Pascale
|
process
|
obsolete go regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication dear all the proposal has been made to obsolete go regulation of basement membrane disassembly involved in semicircular canal fusion by cell communication the reason for obsoletion is that this represents a precomposed term there are no annotations or mappings to this term this term is not present in any subsets you can comment on the ticket thanks pascale
| 1
|
649,239
| 21,260,177,239
|
IssuesEvent
|
2022-04-13 02:43:13
|
apache/dolphinscheduler
|
https://api.github.com/repos/apache/dolphinscheduler
|
closed
|
[Bug-RD][UI Next][V1.0.0-Alpha]The output parameters should not be entered in the task
|
bug need to verify priority:low
|
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened
The output parameters should not be entered in the task
http://ds2:12345/dolphinscheduler/ui/projects/5092091953472/workflow/instances/9925?code=5092095909952
<img width="1919" alt="image" src="https://user-images.githubusercontent.com/76080484/161911825-1e12c1b6-b43e-4b23-a627-0413898a720c.png">
<img width="1793" alt="image" src="https://user-images.githubusercontent.com/76080484/161912157-49d97e60-11aa-4070-b0eb-79b83f424467.png">
### What you expected to happen
If the parameter type is in, the task can be passed in; if the parameter type is out, the task cannot be passed in
### How to reproduce
When the input parameter type is out, it is consistent with the parameter type being in, which is wrong
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
1.0
|
[Bug-RD][UI Next][V1.0.0-Alpha]The output parameters should not be entered in the task - ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened
The output parameters should not be entered in the task
http://ds2:12345/dolphinscheduler/ui/projects/5092091953472/workflow/instances/9925?code=5092095909952
<img width="1919" alt="image" src="https://user-images.githubusercontent.com/76080484/161911825-1e12c1b6-b43e-4b23-a627-0413898a720c.png">
<img width="1793" alt="image" src="https://user-images.githubusercontent.com/76080484/161912157-49d97e60-11aa-4070-b0eb-79b83f424467.png">
### What you expected to happen
If the parameter type is in, the task can be passed in; if the parameter type is out, the task cannot be passed in
### How to reproduce
When the input parameter type is out, it is consistent with the parameter type being in, which is wrong
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
non_process
|
the output parameters should not be entered in the task search before asking i had searched in the and found no similar issues what happened the output parameters should not be entered in the task img width alt image src img width alt image src what you expected to happen if the parameter type is in the task can be passed in if the parameter type is out the task cannot be passed in how to reproduce when the input parameter type is out it is consistent with the parameter type being in which is wrong anything else no response version dev are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
| 0
|
229,058
| 17,500,537,838
|
IssuesEvent
|
2021-08-10 08:54:10
|
imAsparky/sphinxclasstocr
|
https://api.github.com/repos/imAsparky/sphinxclasstocr
|
opened
|
[DOCS]:Update README to reflect this project.
|
documentation
|
**Describe alternatives you've considered**
Would you please share your clear and concise description of any alternative/similar documentation you've considered.
### The README file is misleading because it still points to autoclasstoc.
Update the README to reflect the new project and indicate where it has been forked from.
**Additional context**
Would you please share any other context or screenshots about the documentation request here.
|
1.0
|
[DOCS]:Update README to reflect this project. - **Describe alternatives you've considered**
Would you please share your clear and concise description of any alternative/similar documentation you've considered.
### The README file is misleading because it still points to autoclasstoc.
Update the README to reflect the new project and indicate where it has been forked from.
**Additional context**
Would you please share any other context or screenshots about the documentation request here.
|
non_process
|
update readme to reflect this project describe alternatives you ve considered would you please share your clear and concise description of any alternative similar documentation you ve considered the readme file is misleading because it still points to autoclasstoc update the readme to reflect the new project and indicate where it has been forked from additional context would you please share any other context or screenshots about the documentation request here
| 0
|
20,835
| 27,601,711,342
|
IssuesEvent
|
2023-03-09 10:23:04
|
atc0005/check-process
|
https://api.github.com/repos/atc0005/check-process
|
opened
|
Add support for detecting files with a 0 link count
|
enhancement App: check_process App: lsps linux
|
## Overview
From https://github.com/lsof-org/lsof/issues/65#issuecomment-509019136:
> >a. Finding an Unlinked Open File
> >
> >A pesky variant of a file that is filling a file system is an
> >unlinked file to which some process is still writing. When a
> >process opens a file and then unlinks it, the file's resources
> >remain in use by the process, but the file's directory entries
> >are removed. Hence, even when you know the directory where the
> >file once resided, you can't detect it with ls.
> >
> >This can be an administrative problem when the unlinked file is
> >large, and the process that holds it open continues to write to
> >it. Only when the process closes the file will its resources,
> >particularly disk space, be released.
> >
> >Lsof can help you find unlinked files on local disks. It has an
> >option, +L, that will list the link counts of open files. That
> >helps because an unlinked file on a local disk has a zero link
> >count. Note: this is NOT true for NFS files, accessed from a
> >remote server.
> >
> > You could use the option to list all files and look for a zero
> > link count in the NLINK column -- e.g.,
> >
> > ```
> > $lsof +L
> > COMMAND PID USER FD TYPE DEVICE SIZE/OFF NLINK NODE NAME
> > ...
> > less 25366 abe txt VREG 6,0 40960 1 76319 /usr/...
> > ...
> > less 25366 abe 3r VREG 6,0 17360 0 98768 / (/dev/sd0a)
> > ```
> >
> > Better yet, you can specify an upper bound to the +L option, and
> > lsof will select only files that have a link count less than the
> > upper bound. For example:
> >
> > ```
> > $ lsof +L1
> > COMMAND PID USER FD TYPE DEVICE SIZE/OFF NLINK NODE NAME
> > less 25366 abe 3r VREG 6,0 17360 0 98768 / (/dev/sd0a)
> > ```
> >
> > You can use lsof's -a (AND) option to narrow the link count search
> > to a particular file system. For example, to look for zero link
> > counts on the /home file system, use:
> >
> > ```
> > $ lsof -a +L1 /home
> > ```
> >
> >
> But I don't understood why I'm getting so big difference between lsof's output(for eg. lsof -a +L1) and output of my script which parse every PID stored in procfs. I mean the difference in count of unlinked files.
and the reply at https://github.com/lsof-org/lsof/issues/65#issuecomment-509021151:
> How about `lsof -K -a +L1` ?
## References
- https://github.com/lsof-org/lsof/issues/65
- https://lsof.readthedocs.io/en/latest/tutorial/#finding-an-unlinked-open-file
- https://unix.stackexchange.com/questions/68523/find-and-remove-large-files-that-are-open-but-have-been-deleted
|
1.0
|
Add support for detecting files with a 0 link count - ## Overview
From https://github.com/lsof-org/lsof/issues/65#issuecomment-509019136:
> >a. Finding an Unlinked Open File
> >
> >A pesky variant of a file that is filling a file system is an
> >unlinked file to which some process is still writing. When a
> >process opens a file and then unlinks it, the file's resources
> >remain in use by the process, but the file's directory entries
> >are removed. Hence, even when you know the directory where the
> >file once resided, you can't detect it with ls.
> >
> >This can be an administrative problem when the unlinked file is
> >large, and the process that holds it open continues to write to
> >it. Only when the process closes the file will its resources,
> >particularly disk space, be released.
> >
> >Lsof can help you find unlinked files on local disks. It has an
> >option, +L, that will list the link counts of open files. That
> >helps because an unlinked file on a local disk has a zero link
> >count. Note: this is NOT true for NFS files, accessed from a
> >remote server.
> >
> > You could use the option to list all files and look for a zero
> > link count in the NLINK column -- e.g.,
> >
> > ```
> > $lsof +L
> > COMMAND PID USER FD TYPE DEVICE SIZE/OFF NLINK NODE NAME
> > ...
> > less 25366 abe txt VREG 6,0 40960 1 76319 /usr/...
> > ...
> > less 25366 abe 3r VREG 6,0 17360 0 98768 / (/dev/sd0a)
> > ```
> >
> > Better yet, you can specify an upper bound to the +L option, and
> > lsof will select only files that have a link count less than the
> > upper bound. For example:
> >
> > ```
> > $ lsof +L1
> > COMMAND PID USER FD TYPE DEVICE SIZE/OFF NLINK NODE NAME
> > less 25366 abe 3r VREG 6,0 17360 0 98768 / (/dev/sd0a)
> > ```
> >
> > You can use lsof's -a (AND) option to narrow the link count search
> > to a particular file system. For example, to look for zero link
> > counts on the /home file system, use:
> >
> > ```
> > $ lsof -a +L1 /home
> > ```
> >
> >
> But I don't understood why I'm getting so big difference between lsof's output(for eg. lsof -a +L1) and output of my script which parse every PID stored in procfs. I mean the difference in count of unlinked files.
and the reply at https://github.com/lsof-org/lsof/issues/65#issuecomment-509021151:
> How about `lsof -K -a +L1` ?
## References
- https://github.com/lsof-org/lsof/issues/65
- https://lsof.readthedocs.io/en/latest/tutorial/#finding-an-unlinked-open-file
- https://unix.stackexchange.com/questions/68523/find-and-remove-large-files-that-are-open-but-have-been-deleted
|
process
|
add support for detecting files with a link count overview from a finding an unlinked open file a pesky variant of a file that is filling a file system is an unlinked file to which some process is still writing when a process opens a file and then unlinks it the file s resources remain in use by the process but the file s directory entries are removed hence even when you know the directory where the file once resided you can t detect it with ls this can be an administrative problem when the unlinked file is large and the process that holds it open continues to write to it only when the process closes the file will its resources particularly disk space be released lsof can help you find unlinked files on local disks it has an option l that will list the link counts of open files that helps because an unlinked file on a local disk has a zero link count note this is not true for nfs files accessed from a remote server you could use the option to list all files and look for a zero link count in the nlink column e g lsof l command pid user fd type device size off nlink node name less abe txt vreg usr less abe vreg dev better yet you can specify an upper bound to the l option and lsof will select only files that have a link count less than the upper bound for example lsof command pid user fd type device size off nlink node name less abe vreg dev you can use lsof s a and option to narrow the link count search to a particular file system for example to look for zero link counts on the home file system use lsof a home but i don t understood why i m getting so big difference between lsof s output for eg lsof a and output of my script which parse every pid stored in procfs i mean the difference in count of unlinked files and the reply at how about lsof k a references
| 1
|
15,446
| 19,662,409,758
|
IssuesEvent
|
2022-01-10 18:25:31
|
bridgetownrb/bridgetown
|
https://api.github.com/repos/bridgetownrb/bridgetown
|
closed
|
feat: add the option to use esbuild
|
enhancement process high priority
|
**Update by @jaredcwhite:** focusing on `esbuild` initially, see below…
----
Add a flag to use Snowpack and possibly other bundlers instead of webpack.
`bridgetown new <dir> --bundler=snowpack`
I'm thinking something like the syntax above.
Thoughts?
|
1.0
|
feat: add the option to use esbuild - **Update by @jaredcwhite:** focusing on `esbuild` initially, see below…
----
Add a flag to use Snowpack and possibly other bundlers instead of webpack.
`bridgetown new <dir> --bundler=snowpack`
I'm thinking something like the syntax above.
Thoughts?
|
process
|
feat add the option to use esbuild update by jaredcwhite focusing on esbuild initially see below… add a flag to use snowpack and possibly other bundlers instead of webpack bridgetown new bundler snowpack i m thinking something like the syntax above thoughts
| 1
|
20,307
| 26,947,648,895
|
IssuesEvent
|
2023-02-08 09:21:49
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
SIGSEGV handler causes infinite loop if signal raised from within
|
process lib / src
|
```js
process.on('SIGSEGV', (s) => {
console.log(`program crashed with ${signum}`)
})
// some code that causes node to hard crash, not abort (such as *(int *) 0 = 0;)
```
this causes node to loop tight with high CPU. On the first look I was thinking it keeps calling the handler but that was not the case.
in libuv, the signals (internal or external) are intercepted in `uv__signal_handler` .
- After all the processing, it returns from the signal stack, which causes the faulty instruction to be re-entered, leading to the tight loop.
- it writes to the signal's pipefd, but the thread is never able to reach there to process it and invoke the callback.
These questions come to my mind:
- Is it meaningful to resume to the execution context on internal fatal signals?
- If yes, where should it resume?
- If no, what it should do?
IMO the embedder should decide / control these?
@nodejs/libuv @nodejs/process
|
1.0
|
SIGSEGV handler causes infinite loop if signal raised from within - ```js
process.on('SIGSEGV', (s) => {
console.log(`program crashed with ${signum}`)
})
// some code that causes node to hard crash, not abort (such as *(int *) 0 = 0;)
```
this causes node to loop tight with high CPU. On the first look I was thinking it keeps calling the handler but that was not the case.
in libuv, the signals (internal or external) are intercepted in `uv__signal_handler` .
- After all the processing, it returns from the signal stack, which causes the faulty instruction to be re-entered, leading to the tight loop.
- it writes to the signal's pipefd, but the thread is never able to reach there to process it and invoke the callback.
These questions come to my mind:
- Is it meaningful to resume to the execution context on internal fatal signals?
- If yes, where should it resume?
- If no, what it should do?
IMO the embedder should decide / control these?
@nodejs/libuv @nodejs/process
|
process
|
sigsegv handler causes infinite loop if signal raised from within js process on sigsegv s console log program crashed with signum some code that causes node to hard crash not abort such as int this causes node to loop tight with high cpu on the first look i was thinking it keeps calling the handler but that was not the case in libuv the signals internal or external are intercepted in uv signal handler after all the processing it returns from the signal stack which causes the faulty instruction to be re entered leading to the tight loop it writes to the signal s pipefd but the thread is never able to reach there to process it and invoke the callback these questions come to my mind is it meaningful to resume to the execution context on internal fatal signals if yes where should it resume if no what it should do imo the embedder should decide control these nodejs libuv nodejs process
| 1
|
459,808
| 13,199,604,951
|
IssuesEvent
|
2020-08-14 06:19:25
|
opencrvs/opencrvs-core
|
https://api.github.com/repos/opencrvs/opencrvs-core
|
closed
|
If we don't provide country code before phone number the search option can't find the application.
|
Priority: high 👹Bug
|
**Steps To Reproduce**
1. Log in as field agent.
2. In search field select phone number > provide applicant phone number without country code and search.
3. No result found.
4. Now, search with country code +phone number.
5. Result found.
**Expected behavior**
It should be flexible to search with phone number with/without country code.
**Screenshot**

**Desktop:**
- OS: Windows 10
- Browser: Chrome 84.0, FF 79.0
|
1.0
|
If we don't provide country code before phone number the search option can't find the application. - **Steps To Reproduce**
1. Log in as field agent.
2. In search field select phone number > provide applicant phone number without country code and search.
3. No result found.
4. Now, search with country code +phone number.
5. Result found.
**Expected behavior**
It should be flexible to search with phone number with/without country code.
**Screenshot**

**Desktop:**
- OS: Windows 10
- Browser: Chrome 84.0, FF 79.0
|
non_process
|
if we don t provide country code before phone number the search option can t find the application steps to reproduce log in as field agent in search field select phone number provide applicant phone number without country code and search no result found now search with country code phone number result found expected behavior it should be flexible to search with phone number with without country code screenshot desktop os windows browser chrome ff
| 0
|
20,296
| 26,933,368,383
|
IssuesEvent
|
2023-02-07 18:30:50
|
scverse/anndata
|
https://api.github.com/repos/scverse/anndata
|
opened
|
Remove workaround from test_concat_size_0_dim once upstream bug fixed
|
upstream topic: combining Bug 🐛 dev process
|
* https://github.com/scikit-hep/awkward/issues/2209
Remove marked workaround from `test_concat_size_0_dim` once the upstream bug from awkward is fixed, and a release made.
|
1.0
|
Remove workaround from test_concat_size_0_dim once upstream bug fixed - * https://github.com/scikit-hep/awkward/issues/2209
Remove marked workaround from `test_concat_size_0_dim` once the upstream bug from awkward is fixed, and a release made.
|
process
|
remove workaround from test concat size dim once upstream bug fixed remove marked workaround from test concat size dim once the upstream bug from awkward is fixed and a release made
| 1
|
16,840
| 22,089,025,808
|
IssuesEvent
|
2022-06-01 03:20:25
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Vector Extract By Location not including all features on SQL Server based layers
|
Feedback stale Processing Regression Bug
|
### What is the bug or the crash?
We have a client who has reported what looks like to be a regression between 3.16.13 and 3.22 (and nightly). When using the Extract By Location tool in the processing toolbox the result layer is empty or contains only a single feature when running against SQL Server layers
Using SQL Server-based layers in 3.16.13 will show progress and create a new file with the correct intersected geometries. However, in 3.22.15+ and the latest nightly, it will run really fast and only return a single intersected geometry.
## 3.16.13/.15
These are the options I have used in 3.16.13 (and .15)

with the result being
```
QGIS version: 3.16.13-Hannover
QGIS code revision: a8618a94
Qt version: 5.15.2
GDAL version: 3.3.3
GEOS version: 3.10.0-CAPI-1.16.0
PROJ version: Rel. 8.2.0, November 1st, 2021
Processing algorithm…
Algorithm 'Extract by location' starting…
Input parameters:
{ 'INPUT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiLineString disableInvalidGeometryHandling=\'0\' table=\"dbo\".\"water_pipe_lne\" (geom)', 'INTERSECT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiPolygon disableInvalidGeometryHandling=\'0\' table=\"dbo\".\"Urban_segment_reg\" (geom)', 'OUTPUT' : 'TEMPORARY_OUTPUT', 'PREDICATE' : [0] }
Execution completed in 14.42 seconds
Results:
{'OUTPUT': 'Extracted__location__980c9309_e228_4fac_a05c_295e878e4818'}
Loading resulting layers
Algorithm 'Extract by location' finished
```
Red being the extracted

Layer details:

## 3.22.5
```
QGIS version: 3.22.5-Białowieża
QGIS code revision: c2723178
Qt version: 5.15.2
Python version: 3.9.5
GDAL version: 3.4.1
GEOS version: 3.10.2-CAPI-1.16.0
PROJ version: Rel. 8.2.1, January 1st, 2022
PDAL version: 2.3.0 (git-version: 9f35b7)
Algorithm started at: 2022-04-19T09:17:02
Algorithm 'Extract by location' starting…
Input parameters:
{ 'INPUT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiLineString disableInvalidGeometryHandling=\'0\' table="dbo"."water_pipe_lne" (geom)', 'INTERSECT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiPolygon disableInvalidGeometryHandling=\'0\' table="dbo"."Urban_segment_reg" (geom)', 'OUTPUT' : 'TEMPORARY_OUTPUT', 'PREDICATE' : [0] }
Execution completed in 0.20 seconds
Results:
{'OUTPUT': 'Extracted__location__f11eeee8_1b56_4ca3_b805_500a9876c625'}
Loading resulting layers
Algorithm 'Extract by location' finished
```
Seems to be stopping at the first one:

Layer details of source layers:

## Projection Details
Both layers are reported as GDA94 / MGA zone 55 in both versions. However, the project in both versions seems to report `Unknown CRS`. It looks like a projection issue but strange why it works in 3.16.13 and not latest.
## Works for shapefile
The original source shapefile for both datasets works correctly in both versions, making me think it's something SQL Server related however I couldn't see where this might be an issue in that code.
I have looked and compared the processing code between 3.16 and 3.22 and master but there are no major changes to indicate this would be an issue for only SQL Server as that code is generic.
## Version breakdown
- 3.16.13 - 🟢
- 3.16.15 - 🟢 (seems to work for me on this version)
- 3.22.5 - 🔴
- Nightly - 🔴
### Steps to reproduce the issue
I will work on creating a sample set of data that can be used to reproduce it.
So far it happens with SQL Server layers only.
### Versions
## Working 3.16 version:
QGIS version
3.16.13-Hannover
QGIS code revision
a8618a94
Compiled against Qt
5.15.2
Running against Qt
5.15.2
Compiled against GDAL/OGR
3.3.3
Running against GDAL/OGR
3.3.3
Compiled against GEOS
3.10.0-CAPI-1.16.0
Running against GEOS
3.10.0-CAPI-1.16.0
Compiled against SQLite
3.35.2
Running against SQLite
3.35.2
PostgreSQL Client Version
13.0
SpatiaLite Version
5.0.1
QWT Version
6.1.3
QScintilla2 Version
2.11.5
Compiled against PROJ
8.2.0
Running against PROJ
Rel. 8.2.0, November 1st, 2021
OS Version
Windows 10 Version 2009
Active python plugins
BoundingBox;
chartis_assets;
grid_reference_locator-master;
latlontools;
plugin_reloader;
db_manager;
MetaSearch;
processing
## 3.22
QGIS version
3.22.5-Białowieża
QGIS code revision
c2723178
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.1
PROJ version
8.2.1
EPSG Registry database version
v10.041 (2021-12-03)
GEOS version
3.10.2-CAPI-1.16.0
SQLite version
3.37.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
BoundingBox
2.1
chartis_assets
0.0.1
grid_reference_locator-master
0.1
latlontools
3.3.17
plugin_reloader
0.9.1
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
Vector Extract By Location not including all features on SQL Server based layers - ### What is the bug or the crash?
We have a client who has reported what looks like to be a regression between 3.16.13 and 3.22 (and nightly). When using the Extract By Location tool in the processing toolbox the result layer is empty or contains only a single feature when running against SQL Server layers
Using SQL Server-based layers in 3.16.13 will show progress and create a new file with the correct intersected geometries. However, in 3.22.15+ and the latest nightly, it will run really fast and only return a single intersected geometry.
## 3.16.13/.15
These are the options I have used in 3.16.13 (and .15)

with the result being
```
QGIS version: 3.16.13-Hannover
QGIS code revision: a8618a94
Qt version: 5.15.2
GDAL version: 3.3.3
GEOS version: 3.10.0-CAPI-1.16.0
PROJ version: Rel. 8.2.0, November 1st, 2021
Processing algorithm…
Algorithm 'Extract by location' starting…
Input parameters:
{ 'INPUT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiLineString disableInvalidGeometryHandling=\'0\' table=\"dbo\".\"water_pipe_lne\" (geom)', 'INTERSECT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiPolygon disableInvalidGeometryHandling=\'0\' table=\"dbo\".\"Urban_segment_reg\" (geom)', 'OUTPUT' : 'TEMPORARY_OUTPUT', 'PREDICATE' : [0] }
Execution completed in 14.42 seconds
Results:
{'OUTPUT': 'Extracted__location__980c9309_e228_4fac_a05c_295e878e4818'}
Loading resulting layers
Algorithm 'Extract by location' finished
```
Red being the extracted

Layer details:

## 3.22.5
```
QGIS version: 3.22.5-Białowieża
QGIS code revision: c2723178
Qt version: 5.15.2
Python version: 3.9.5
GDAL version: 3.4.1
GEOS version: 3.10.2-CAPI-1.16.0
PROJ version: Rel. 8.2.1, January 1st, 2022
PDAL version: 2.3.0 (git-version: 9f35b7)
Algorithm started at: 2022-04-19T09:17:02
Algorithm 'Extract by location' starting…
Input parameters:
{ 'INPUT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiLineString disableInvalidGeometryHandling=\'0\' table="dbo"."water_pipe_lne" (geom)', 'INTERSECT' : 'mssql://dbname=\'wat\' host=localhost estimatedmetadata=true srid=28355 type=MultiPolygon disableInvalidGeometryHandling=\'0\' table="dbo"."Urban_segment_reg" (geom)', 'OUTPUT' : 'TEMPORARY_OUTPUT', 'PREDICATE' : [0] }
Execution completed in 0.20 seconds
Results:
{'OUTPUT': 'Extracted__location__f11eeee8_1b56_4ca3_b805_500a9876c625'}
Loading resulting layers
Algorithm 'Extract by location' finished
```
Seems to be stopping at the first one:

Layer details of source layers:

## Projection Details
Both layers are reported as GDA94 / MGA zone 55 in both versions. However, the project in both versions seems to report `Unknown CRS`. It looks like a projection issue but strange why it works in 3.16.13 and not latest.
## Works for shapefile
The original source shapefile for both datasets works correctly in both versions, making me think it's something SQL Server related however I couldn't see where this might be an issue in that code.
I have looked and compared the processing code between 3.16 and 3.22 and master but there are no major changes to indicate this would be an issue for only SQL Server as that code is generic.
## Version breakdown
- 3.16.13 - 🟢
- 3.16.15 - 🟢 (seems to work for me on this version)
- 3.22.5 - 🔴
- Nightly - 🔴
### Steps to reproduce the issue
I will work on creating a sample set of data that can be used to reproduce it.
So far it happens with SQL Server layers only.
### Versions
## Working 3.16 version:
QGIS version
3.16.13-Hannover
QGIS code revision
a8618a94
Compiled against Qt
5.15.2
Running against Qt
5.15.2
Compiled against GDAL/OGR
3.3.3
Running against GDAL/OGR
3.3.3
Compiled against GEOS
3.10.0-CAPI-1.16.0
Running against GEOS
3.10.0-CAPI-1.16.0
Compiled against SQLite
3.35.2
Running against SQLite
3.35.2
PostgreSQL Client Version
13.0
SpatiaLite Version
5.0.1
QWT Version
6.1.3
QScintilla2 Version
2.11.5
Compiled against PROJ
8.2.0
Running against PROJ
Rel. 8.2.0, November 1st, 2021
OS Version
Windows 10 Version 2009
Active python plugins
BoundingBox;
chartis_assets;
grid_reference_locator-master;
latlontools;
plugin_reloader;
db_manager;
MetaSearch;
processing
## 3.22
QGIS version
3.22.5-Białowieża
QGIS code revision
c2723178
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.1
PROJ version
8.2.1
EPSG Registry database version
v10.041 (2021-12-03)
GEOS version
3.10.2-CAPI-1.16.0
SQLite version
3.37.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
BoundingBox
2.1
chartis_assets
0.0.1
grid_reference_locator-master
0.1
latlontools
3.3.17
plugin_reloader
0.9.1
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
vector extract by location not including all features on sql server based layers what is the bug or the crash we have a client who has reported what looks like to be a regression between and and nightly when using the extract by location tool in the processing toolbox the result layer is empty or contains only a single feature when running against sql server layers using sql server based layers in will show progress and create a new file with the correct intersected geometries however in and the latest nightly it will run really fast and only return a single intersected geometry these are the options i have used in and with the result being qgis version hannover qgis code revision qt version gdal version geos version capi proj version rel november processing algorithm… algorithm extract by location starting… input parameters input mssql dbname wat host localhost estimatedmetadata true srid type multilinestring disableinvalidgeometryhandling table dbo water pipe lne geom intersect mssql dbname wat host localhost estimatedmetadata true srid type multipolygon disableinvalidgeometryhandling table dbo urban segment reg geom output temporary output predicate execution completed in seconds results output extracted location loading resulting layers algorithm extract by location finished red being the extracted layer details qgis version białowieża qgis code revision qt version python version gdal version geos version capi proj version rel january pdal version git version algorithm started at algorithm extract by location starting… input parameters input mssql dbname wat host localhost estimatedmetadata true srid type multilinestring disableinvalidgeometryhandling table dbo water pipe lne geom intersect mssql dbname wat host localhost estimatedmetadata true srid type multipolygon disableinvalidgeometryhandling table dbo urban segment reg geom output temporary output predicate execution completed in seconds results output extracted location loading resulting layers algorithm extract by location finished seems to be stopping at the first one layer details of source layers projection details both layers are reported as mga zone in both versions however the project in both versions seems to report unknown crs it looks like a projection issue but strange why it works in and not latest works for shapefile the original source shapefile for both datasets works correctly in both versions making me think it s something sql server related however i couldn t see where this might be an issue in that code i have looked and compared the processing code between and and master but there are no major changes to indicate this would be an issue for only sql server as that code is generic version breakdown 🟢 🟢 seems to work for me on this version 🔴 nightly 🔴 steps to reproduce the issue i will work on creating a sample set of data that can be used to reproduce it so far it happens with sql server layers only versions working version qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel november os version windows version active python plugins boundingbox chartis assets grid reference locator master latlontools plugin reloader db manager metasearch processing qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins boundingbox chartis assets grid reference locator master latlontools plugin reloader db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
527,823
| 15,353,395,094
|
IssuesEvent
|
2021-03-01 08:31:00
|
Conjurinc-workato-dev/evoke
|
https://api.github.com/repos/Conjurinc-workato-dev/evoke
|
reopened
|
testing sorting Integration User out
|
ONYX-6632 kind/bug priority/Major team/Palmtree triage/Cannot Reproduce
|
##description
Modified issue Description from GitHub to Jira
##
|
1.0
|
testing sorting Integration User out - ##description
Modified issue Description from GitHub to Jira
##
|
non_process
|
testing sorting integration user out description modified issue description from github to jira
| 0
|
13,886
| 8,392,442,712
|
IssuesEvent
|
2018-10-09 17:38:41
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
sql: improve precision of index spans to scan for multi-column-family single-row fetches
|
A-sql-planning C-performance
|
Suppose we have a table and a query:
```
CREATE TABLE a (a INT PRIMARY KEY, b INT, c INT, FAMILY(a), FAMILY(b), FAMILY(c);
INSERT INTO a VALUES(1,2,3);
SELECT b FROM a WHERE a=1;
```
Currently, this query will cause a scan from `/a/primary/1 -> /a/primary/2`, even though ideally we'd only need to scan the key at `/a/primary/1/b`.
We have enough information to improve this scan, since we know that `c` is not a needed column, `a` is the primary index and therefore unique, and we have an equality constraint on `a`.
We should consider improving the algorithm in `spanFromLogicalSpan` to incorporate all of this information into its final output span.
cc @danhhz @andreimatei @petermattis based on our earlier conversation.
|
True
|
sql: improve precision of index spans to scan for multi-column-family single-row fetches - Suppose we have a table and a query:
```
CREATE TABLE a (a INT PRIMARY KEY, b INT, c INT, FAMILY(a), FAMILY(b), FAMILY(c);
INSERT INTO a VALUES(1,2,3);
SELECT b FROM a WHERE a=1;
```
Currently, this query will cause a scan from `/a/primary/1 -> /a/primary/2`, even though ideally we'd only need to scan the key at `/a/primary/1/b`.
We have enough information to improve this scan, since we know that `c` is not a needed column, `a` is the primary index and therefore unique, and we have an equality constraint on `a`.
We should consider improving the algorithm in `spanFromLogicalSpan` to incorporate all of this information into its final output span.
cc @danhhz @andreimatei @petermattis based on our earlier conversation.
|
non_process
|
sql improve precision of index spans to scan for multi column family single row fetches suppose we have a table and a query create table a a int primary key b int c int family a family b family c insert into a values select b from a where a currently this query will cause a scan from a primary a primary even though ideally we d only need to scan the key at a primary b we have enough information to improve this scan since we know that c is not a needed column a is the primary index and therefore unique and we have an equality constraint on a we should consider improving the algorithm in spanfromlogicalspan to incorporate all of this information into its final output span cc danhhz andreimatei petermattis based on our earlier conversation
| 0
|
772,761
| 27,134,586,497
|
IssuesEvent
|
2023-02-16 12:18:19
|
cpp-lln-lab/bidspm
|
https://api.github.com/repos/cpp-lln-lab/bidspm
|
closed
|
wget requested to download demo dataset
|
high priority
|
When running `demos/test_moae.m`, the following messages appear on the command window:
> zsh:1: command not found: wget
followed by the error
> Error using movefile
> No matching files named '/../bidspm/demos/MoAE/MoAEpilot.bids.zip' were found.
EDIT:
`system` within matlab requires the full path of the command to run (e.g. `wget`) on Mac.
It's obtainable running `where wget` on a terminal and updating the path in the demo.
For reference: https://nl.mathworks.com/matlabcentral/answers/740062-use-command-line-in-matlab-using-and-zsh-1-command-not-found-is-displayed
|
1.0
|
wget requested to download demo dataset - When running `demos/test_moae.m`, the following messages appear on the command window:
> zsh:1: command not found: wget
followed by the error
> Error using movefile
> No matching files named '/../bidspm/demos/MoAE/MoAEpilot.bids.zip' were found.
EDIT:
`system` within matlab requires the full path of the command to run (e.g. `wget`) on Mac.
It's obtainable running `where wget` on a terminal and updating the path in the demo.
For reference: https://nl.mathworks.com/matlabcentral/answers/740062-use-command-line-in-matlab-using-and-zsh-1-command-not-found-is-displayed
|
non_process
|
wget requested to download demo dataset when running demos test moae m the following messages appear on the command window zsh command not found wget followed by the error error using movefile no matching files named bidspm demos moae moaepilot bids zip were found edit system within matlab requires the full path of the command to run e g wget on mac it s obtainable running where wget on a terminal and updating the path in the demo for reference
| 0
|
705,394
| 24,233,286,532
|
IssuesEvent
|
2022-09-26 20:19:50
|
djordjemoj/s2s
|
https://api.github.com/repos/djordjemoj/s2s
|
closed
|
[Prijava] Broj telefona ne sme biti negativan niti u losem formatu
|
priority
|
Na prijavi moze da se prijavi sa negativnim brojem, izbaciti one strelice za biranje brojeva
|
1.0
|
[Prijava] Broj telefona ne sme biti negativan niti u losem formatu - Na prijavi moze da se prijavi sa negativnim brojem, izbaciti one strelice za biranje brojeva
|
non_process
|
broj telefona ne sme biti negativan niti u losem formatu na prijavi moze da se prijavi sa negativnim brojem izbaciti one strelice za biranje brojeva
| 0
|
4,351
| 7,254,899,308
|
IssuesEvent
|
2018-02-16 12:57:52
|
ContaoMonitoring/monitoring-tasks
|
https://api.github.com/repos/ContaoMonitoring/monitoring-tasks
|
closed
|
Fix NOT NULL field without default value
|
Defect ⚙ - Processed
|
For further information have a look at: ContaoMonitoring/monitoring#18
- https://github.com/ContaoMonitoring/monitoring-tasks/blob/master/system/modules/MonitoringTasks/dca/tl_monitoring_task.php#L214
- https://github.com/ContaoMonitoring/monitoring-tasks/blob/master/system/modules/MonitoringTasks/dca/tl_monitoring_task.php#L226
|
1.0
|
Fix NOT NULL field without default value - For further information have a look at: ContaoMonitoring/monitoring#18
- https://github.com/ContaoMonitoring/monitoring-tasks/blob/master/system/modules/MonitoringTasks/dca/tl_monitoring_task.php#L214
- https://github.com/ContaoMonitoring/monitoring-tasks/blob/master/system/modules/MonitoringTasks/dca/tl_monitoring_task.php#L226
|
process
|
fix not null field without default value for further information have a look at contaomonitoring monitoring
| 1
|
468
| 2,905,750,516
|
IssuesEvent
|
2015-06-19 03:28:34
|
open-app/holodex
|
https://api.github.com/repos/open-app/holodex
|
opened
|
problem: how do we know how much someone has worked on Holodex?
|
discussion process
|
proposed solution: we [track our hours](https://www.npmjs.com/package/clocker) and put them in a simple csv format somewhere in this repo. bonus points for analyzing the data and drawing pretty graphs, but we can do that later.
might help with the block of #58, also will help with us being a commons-based peer production value network.
|
1.0
|
problem: how do we know how much someone has worked on Holodex? - proposed solution: we [track our hours](https://www.npmjs.com/package/clocker) and put them in a simple csv format somewhere in this repo. bonus points for analyzing the data and drawing pretty graphs, but we can do that later.
might help with the block of #58, also will help with us being a commons-based peer production value network.
|
process
|
problem how do we know how much someone has worked on holodex proposed solution we and put them in a simple csv format somewhere in this repo bonus points for analyzing the data and drawing pretty graphs but we can do that later might help with the block of also will help with us being a commons based peer production value network
| 1
|
14,287
| 17,261,026,698
|
IssuesEvent
|
2021-07-22 07:39:20
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[MacCatalyst][libraries] System.Diagnostics.Process.Tests fails on MacCatalyst
|
area-System.Diagnostics.Process os-maccatalyst
|
System.Diagnostics.Process.Tests.dll Failed: 7
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
```
System.Diagnostics.Process.Tests.dll Failed: 7
Test collection for System.Diagnostics.Process.Tests.dll
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported
System.ComponentModel.Win32Exception : Undefined error: 0
|
1.0
|
[MacCatalyst][libraries] System.Diagnostics.Process.Tests fails on MacCatalyst - System.Diagnostics.Process.Tests.dll Failed: 7
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
```
System.Diagnostics.Process.Tests.dll Failed: 7
Test collection for System.Diagnostics.Process.Tests.dll
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.ComponentModel.Win32Exception : No such file or directory
System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported
System.ComponentModel.Win32Exception : Undefined error: 0
|
process
|
system diagnostics process tests fails on maccatalyst system diagnostics process tests dll failed system diagnostics tests processtests maxworkingset getnotstarted throwsinvalidoperationexception system diagnostics tests processtests processstart useshellexecute executes filenameasurl false system diagnostics tests processtests processstart useshellexecute executes filenameasurl true system diagnostics tests processtests processstart useshellexecute workingdirectory system diagnostics tests processtests processnamematchesscriptname system diagnostics tests processtests longprocessnamesaresupported system diagnostics tests processtests minworkingset getnotstarted throwsinvalidoperationexception system diagnostics process tests dll failed test collection for system diagnostics process tests dll system diagnostics tests processtests processstart useshellexecute executes filenameasurl false system componentmodel no such file or directory system diagnostics tests processtests processstart useshellexecute executes filenameasurl true system componentmodel no such file or directory system diagnostics tests processtests processstart useshellexecute workingdirectory system componentmodel no such file or directory system diagnostics tests processtests processnamematchesscriptname system componentmodel no such file or directory system diagnostics tests processtests longprocessnamesaresupported system componentmodel undefined error
| 1
|
17,517
| 23,329,214,049
|
IssuesEvent
|
2022-08-09 02:11:41
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[Feature][FLINK-28351][Stream API] Pulsar Sink should support dynamic generated topic from record
|
compute/data-processing type/feature
|
Some people would like to use dynamically-generated topics from messages and use the key hash range policy. This is not supported by the Pulsar sink currently. We would introduce a new interface named `TopicExacter` and add a new `setTopics(TopicExacter)` in `PulsarSinkBuilder`.
|
1.0
|
[Feature][FLINK-28351][Stream API] Pulsar Sink should support dynamic generated topic from record - Some people would like to use dynamically-generated topics from messages and use the key hash range policy. This is not supported by the Pulsar sink currently. We would introduce a new interface named `TopicExacter` and add a new `setTopics(TopicExacter)` in `PulsarSinkBuilder`.
|
process
|
pulsar sink should support dynamic generated topic from record some people would like to use dynamically generated topics from messages and use the key hash range policy this is not supported by the pulsar sink currently we would introduce a new interface named topicexacter and add a new settopics topicexacter in pulsarsinkbuilder
| 1
|
315,285
| 23,550,069,845
|
IssuesEvent
|
2022-08-21 17:49:12
|
wb8tyw/D-Rats
|
https://api.github.com/repos/wb8tyw/D-Rats
|
closed
|
Add yamllint to git pre-commit and github actions
|
documentation enhancement Fixed
|
We have yaml files for github actions now, so we should check them.
|
1.0
|
Add yamllint to git pre-commit and github actions - We have yaml files for github actions now, so we should check them.
|
non_process
|
add yamllint to git pre commit and github actions we have yaml files for github actions now so we should check them
| 0
|
14,743
| 18,014,697,548
|
IssuesEvent
|
2021-09-16 12:45:11
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Obsoletion notice: GO:0039647 suppression by virus of host poly(A)-binding protein activity and suppression by virus of host protease activator activity
|
obsoletion multi-species process
|
Dear all,
The proposal has been made to obsolete
GO:0039647 suppression by virus of host poly(A)-binding protein activity
GO:0110087 suppression by virus of host protease activator activity
The reason for obsoletion is that these are molecular functions represented as processes. There are no annotations or mappings to these terms; these terms are not present in any subsets.
You may comment on the ticket:
Thanks, Pascale
|
1.0
|
Obsoletion notice: GO:0039647 suppression by virus of host poly(A)-binding protein activity and suppression by virus of host protease activator activity - Dear all,
The proposal has been made to obsolete
GO:0039647 suppression by virus of host poly(A)-binding protein activity
GO:0110087 suppression by virus of host protease activator activity
The reason for obsoletion is that these are molecular functions represented as processes. There are no annotations or mappings to these terms; these terms are not present in any subsets.
You may comment on the ticket:
Thanks, Pascale
|
process
|
obsoletion notice go suppression by virus of host poly a binding protein activity and suppression by virus of host protease activator activity dear all the proposal has been made to obsolete go suppression by virus of host poly a binding protein activity go suppression by virus of host protease activator activity the reason for obsoletion is that these are molecular functions represented as processes there are no annotations or mappings to these terms these terms are not present in any subsets you may comment on the ticket thanks pascale
| 1
|
5,062
| 7,867,331,872
|
IssuesEvent
|
2018-06-23 07:10:18
|
vtloc/grokking-links
|
https://api.github.com/repos/vtloc/grokking-links
|
opened
|
Trunk Based Development
|
Software Engineering Process
|
Git là công cụ quản lý source code rất phổ biến và dễ dùng hiện nay. Tuy nhiên, mặc dù Git rất dễ sử dụng, việc dùng git một cách hiệu quả là không hề đơn giản tí nào.
Có khá nhiều cách tổ chức Git của dự án của bạn, trong đó có chiến lược GitFlow, Trunk-based cũng như vài cách tổ chức khác.
Đây là một website khá hữu ích tổng hợp thông tin về cách tổ chức nhánh Git theo chiến lược Trunk-based. Mời các bạn cùng đọc.
https://trunkbaseddevelopment.com/
|
1.0
|
Trunk Based Development - Git là công cụ quản lý source code rất phổ biến và dễ dùng hiện nay. Tuy nhiên, mặc dù Git rất dễ sử dụng, việc dùng git một cách hiệu quả là không hề đơn giản tí nào.
Có khá nhiều cách tổ chức Git của dự án của bạn, trong đó có chiến lược GitFlow, Trunk-based cũng như vài cách tổ chức khác.
Đây là một website khá hữu ích tổng hợp thông tin về cách tổ chức nhánh Git theo chiến lược Trunk-based. Mời các bạn cùng đọc.
https://trunkbaseddevelopment.com/
|
process
|
trunk based development git là công cụ quản lý source code rất phổ biến và dễ dùng hiện nay tuy nhiên mặc dù git rất dễ sử dụng việc dùng git một cách hiệu quả là không hề đơn giản tí nào có khá nhiều cách tổ chức git của dự án của bạn trong đó có chiến lược gitflow trunk based cũng như vài cách tổ chức khác đây là một website khá hữu ích tổng hợp thông tin về cách tổ chức nhánh git theo chiến lược trunk based mời các bạn cùng đọc
| 1
|
44,501
| 13,056,855,359
|
IssuesEvent
|
2020-07-30 05:58:49
|
sshivananda/ts-sqs-consumer
|
https://api.github.com/repos/sshivananda/ts-sqs-consumer
|
opened
|
CVE-2020-11023 (Medium) detected in jquery-2.1.3.min.js
|
security vulnerability
|
## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/ts-sqs-consumer/node_modules/knuth-shuffle-seeded/index.html</p>
<p>Path to vulnerable library: /ts-sqs-consumer/node_modules/knuth-shuffle-seeded/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.3.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sshivananda/ts-sqs-consumer/commit/8e86a2adfdf841f4ff57d761e7ba0359998b420d">8e86a2adfdf841f4ff57d761e7ba0359998b420d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11023 (Medium) detected in jquery-2.1.3.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/ts-sqs-consumer/node_modules/knuth-shuffle-seeded/index.html</p>
<p>Path to vulnerable library: /ts-sqs-consumer/node_modules/knuth-shuffle-seeded/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.3.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sshivananda/ts-sqs-consumer/commit/8e86a2adfdf841f4ff57d761e7ba0359998b420d">8e86a2adfdf841f4ff57d761e7ba0359998b420d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm ts sqs consumer node modules knuth shuffle seeded index html path to vulnerable library ts sqs consumer node modules knuth shuffle seeded index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
18,480
| 24,550,740,723
|
IssuesEvent
|
2022-10-12 12:25:24
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] [Angular Upgrade] Password criteria should be aligned as per the design document
|
Bug P1 Participant manager Process: Fixed Process: Tested dev
|
Password criteria should be aligned as per the design document
**Note:** Issue needs to be fixed wherever 'Password criteria' is available
**AR:**

**ER:**

|
2.0
|
[PM] [Angular Upgrade] Password criteria should be aligned as per the design document - Password criteria should be aligned as per the design document
**Note:** Issue needs to be fixed wherever 'Password criteria' is available
**AR:**

**ER:**

|
process
|
password criteria should be aligned as per the design document password criteria should be aligned as per the design document note issue needs to be fixed wherever password criteria is available ar er
| 1
|
36,869
| 12,427,134,134
|
IssuesEvent
|
2020-05-25 01:06:07
|
jgeraigery/uplus-wss
|
https://api.github.com/repos/jgeraigery/uplus-wss
|
opened
|
WS-2019-0424 (Medium) detected in elliptic-6.5.2.tgz
|
security vulnerability
|
## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/uplus-wss/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/uplus-wss/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-babel-4.3.1.tgz (Root Library)
- webpack-4.43.0.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.1.0.tgz
- :x: **elliptic-6.5.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions before 6.5.2 of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.2","isTransitiveDependency":true,"dependencyTree":"@vue/cli-plugin-babel:4.3.1;webpack:4.43.0;node-libs-browser:2.2.1;crypto-browserify:3.12.0;browserify-sign:4.1.0;elliptic:6.5.2","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2019-0424","vulnerabilityDetails":"all versions before 6.5.2 of elliptic are vulnerable to Timing Attack through side-channels.","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Adjacent","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2019-0424 (Medium) detected in elliptic-6.5.2.tgz - ## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/uplus-wss/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/uplus-wss/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-babel-4.3.1.tgz (Root Library)
- webpack-4.43.0.tgz
- node-libs-browser-2.2.1.tgz
- crypto-browserify-3.12.0.tgz
- browserify-sign-4.1.0.tgz
- :x: **elliptic-6.5.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions before 6.5.2 of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.2","isTransitiveDependency":true,"dependencyTree":"@vue/cli-plugin-babel:4.3.1;webpack:4.43.0;node-libs-browser:2.2.1;crypto-browserify:3.12.0;browserify-sign:4.1.0;elliptic:6.5.2","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2019-0424","vulnerabilityDetails":"all versions before 6.5.2 of elliptic are vulnerable to Timing Attack through side-channels.","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Adjacent","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file tmp ws scm uplus wss package json path to vulnerable library tmp ws scm uplus wss node modules elliptic package json dependency hierarchy cli plugin babel tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library vulnerability details all versions before of elliptic are vulnerable to timing attack through side channels publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact none for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails all versions before of elliptic are vulnerable to timing attack through side channels vulnerabilityurl
| 0
|
7,230
| 10,368,891,128
|
IssuesEvent
|
2019-09-07 20:50:05
|
kubeflow/testing
|
https://api.github.com/repos/kubeflow/testing
|
closed
|
Standardize/cleanup different bug kind labels
|
area/engprod good first issue help wanted kind/process priority/p1
|
The following labels are currently used to indicate what kind an issue is
* community/discussion
* community/question
* kind/bug
* improvement/enhancement
* addition/feature
* kind/process
Not using the kind prefix has a bunch of drawbacks
* Prow allows the "/kind" command to be used to set the kind.
* We use this in our issue templates to automatically set the kind
* If we don't use kind as a prefix people can't set it via the kind bot
* It makes our [issue triage process](https://github.com/kubeflow/community/blob/master/proposals/issue_triage.md) more complicated because we'd like to set different requirements based on the kind
* e.g. A discussion probably doesn't need a priority or be assigned to a project
I propose we use the following kinds
* kind/bug
* kind/question
* kind/feature
* kind/discussion
* kind/process
I think we want to make the following changes
* Update [kubeflow_label.yml](https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml)
* Map the old times to the new types
* Update the running instance of label sync and make sure it succesfully applies across all repos
* Update the issue label bot configs ([example](https://github.com/kubeflow/kubeflow/blob/master/.github/issue_label_bot.yaml)) in all repos so that it uses the new kinds
|
1.0
|
Standardize/cleanup different bug kind labels - The following labels are currently used to indicate what kind an issue is
* community/discussion
* community/question
* kind/bug
* improvement/enhancement
* addition/feature
* kind/process
Not using the kind prefix has a bunch of drawbacks
* Prow allows the "/kind" command to be used to set the kind.
* We use this in our issue templates to automatically set the kind
* If we don't use kind as a prefix people can't set it via the kind bot
* It makes our [issue triage process](https://github.com/kubeflow/community/blob/master/proposals/issue_triage.md) more complicated because we'd like to set different requirements based on the kind
* e.g. A discussion probably doesn't need a priority or be assigned to a project
I propose we use the following kinds
* kind/bug
* kind/question
* kind/feature
* kind/discussion
* kind/process
I think we want to make the following changes
* Update [kubeflow_label.yml](https://github.com/kubeflow/testing/blob/master/label_sync/kubeflow_label.yml)
* Map the old times to the new types
* Update the running instance of label sync and make sure it succesfully applies across all repos
* Update the issue label bot configs ([example](https://github.com/kubeflow/kubeflow/blob/master/.github/issue_label_bot.yaml)) in all repos so that it uses the new kinds
|
process
|
standardize cleanup different bug kind labels the following labels are currently used to indicate what kind an issue is community discussion community question kind bug improvement enhancement addition feature kind process not using the kind prefix has a bunch of drawbacks prow allows the kind command to be used to set the kind we use this in our issue templates to automatically set the kind if we don t use kind as a prefix people can t set it via the kind bot it makes our more complicated because we d like to set different requirements based on the kind e g a discussion probably doesn t need a priority or be assigned to a project i propose we use the following kinds kind bug kind question kind feature kind discussion kind process i think we want to make the following changes update map the old times to the new types update the running instance of label sync and make sure it succesfully applies across all repos update the issue label bot configs in all repos so that it uses the new kinds
| 1
|
16,980
| 22,338,968,855
|
IssuesEvent
|
2022-06-14 21:39:34
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
opened
|
Update PHAST Fixture, Wall and Atmosphere losses when sql db items deleted
|
bug Process Heating
|
Modify PHAST Fixture, Wall and Atmosphere losses forms to store material object properties, so that a custom material can be restored if deleted form the db
|
1.0
|
Update PHAST Fixture, Wall and Atmosphere losses when sql db items deleted - Modify PHAST Fixture, Wall and Atmosphere losses forms to store material object properties, so that a custom material can be restored if deleted form the db
|
process
|
update phast fixture wall and atmosphere losses when sql db items deleted modify phast fixture wall and atmosphere losses forms to store material object properties so that a custom material can be restored if deleted form the db
| 1
|
3,518
| 2,770,853,888
|
IssuesEvent
|
2015-05-01 17:31:22
|
Norhaven/FluentBoilerplate
|
https://api.github.com/repos/Norhaven/FluentBoilerplate
|
opened
|
Create an architectural overview diagram
|
documentation priority: medium up for grabs
|
It would be useful, especially for new people who would like to contribute, to have a good idea of where everything is and how things are broken up into layers and sections. Just need a Visio diagram and an image version that can be embedded into a wiki page.
- Two primary layers (public contract and internal)
- Inside primary layers, namespaces and their purpose
- Inside namespaces, if useful, call out the most commonly used or primary types
|
1.0
|
Create an architectural overview diagram - It would be useful, especially for new people who would like to contribute, to have a good idea of where everything is and how things are broken up into layers and sections. Just need a Visio diagram and an image version that can be embedded into a wiki page.
- Two primary layers (public contract and internal)
- Inside primary layers, namespaces and their purpose
- Inside namespaces, if useful, call out the most commonly used or primary types
|
non_process
|
create an architectural overview diagram it would be useful especially for new people who would like to contribute to have a good idea of where everything is and how things are broken up into layers and sections just need a visio diagram and an image version that can be embedded into a wiki page two primary layers public contract and internal inside primary layers namespaces and their purpose inside namespaces if useful call out the most commonly used or primary types
| 0
|
7,707
| 10,817,322,505
|
IssuesEvent
|
2019-11-08 09:29:13
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
closed
|
Errors running Prisma on GitHub Actions (linux)
|
process/candidate
|
### What I did
Set up a new prisma2 project with the following options:
- Starter Project
- Javascript
- GraphQL + Auth
- SQLite
I then set up CI using Github Actions, to npm install and run the build command.
### Expected Behaviour
The project should build and the `dist` folder generated on completion.
### Describe the bug
`prisma2 generate` fails with:
```bash
$ npm -s run generate
18
Error: Error: Command failed with exit code 2 (ENOENT): /github/workspace/node_modules/prisma2/query-engine-linux-glibc-libssl1.1.0 cli --dmmf
19
spawn /github/workspace/node_modules/prisma2/query-engine-linux-glibc-libssl1.1.0 ENOENT
20
error Command failed with exit code 1.
21
info Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command.
22
##[error]Docker run failed with exit code 1
```
### To Reproduce
I've set up a repo with said options and github actions https://github.com/iRoachie/prisma2-github-ci
---
Prisma Version: `2.0.0-preview015`
|
1.0
|
Errors running Prisma on GitHub Actions (linux) - ### What I did
Set up a new prisma2 project with the following options:
- Starter Project
- Javascript
- GraphQL + Auth
- SQLite
I then set up CI using Github Actions, to npm install and run the build command.
### Expected Behaviour
The project should build and the `dist` folder generated on completion.
### Describe the bug
`prisma2 generate` fails with:
```bash
$ npm -s run generate
18
Error: Error: Command failed with exit code 2 (ENOENT): /github/workspace/node_modules/prisma2/query-engine-linux-glibc-libssl1.1.0 cli --dmmf
19
spawn /github/workspace/node_modules/prisma2/query-engine-linux-glibc-libssl1.1.0 ENOENT
20
error Command failed with exit code 1.
21
info Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command.
22
##[error]Docker run failed with exit code 1
```
### To Reproduce
I've set up a repo with said options and github actions https://github.com/iRoachie/prisma2-github-ci
---
Prisma Version: `2.0.0-preview015`
|
process
|
errors running prisma on github actions linux what i did set up a new project with the following options starter project javascript graphql auth sqlite i then set up ci using github actions to npm install and run the build command expected behaviour the project should build and the dist folder generated on completion describe the bug generate fails with bash npm s run generate error error command failed with exit code enoent github workspace node modules query engine linux glibc cli dmmf spawn github workspace node modules query engine linux glibc enoent error command failed with exit code info visit for documentation about this command docker run failed with exit code to reproduce i ve set up a repo with said options and github actions prisma version
| 1
|
22,078
| 30,597,135,553
|
IssuesEvent
|
2023-07-22 00:20:45
|
aolabNeuro/analyze
|
https://api.github.com/repos/aolabNeuro/analyze
|
opened
|
hand and cursor kinematics in different coordinate systems
|
bug invalid preprocessing
|
interp_cursor and interp_hand are in different coordinate systems in preproc data, which is confusing. they should be updated.
|
1.0
|
hand and cursor kinematics in different coordinate systems - interp_cursor and interp_hand are in different coordinate systems in preproc data, which is confusing. they should be updated.
|
process
|
hand and cursor kinematics in different coordinate systems interp cursor and interp hand are in different coordinate systems in preproc data which is confusing they should be updated
| 1
|
202,680
| 15,295,415,229
|
IssuesEvent
|
2021-02-24 04:47:30
|
Landry333/Big-Owl
|
https://api.github.com/repos/Landry333/Big-Owl
|
closed
|
TEST-124: Local/Instrumented Unit Test for SupervisedGroupList Activity
|
high priority testing
|
Children of Issue[#97]
Testings for SupervisedGroupList and SupervisedGroupPage
- [x] SupervisedGroupList
- [x] SupervisedGroupPage
|
1.0
|
TEST-124: Local/Instrumented Unit Test for SupervisedGroupList Activity - Children of Issue[#97]
Testings for SupervisedGroupList and SupervisedGroupPage
- [x] SupervisedGroupList
- [x] SupervisedGroupPage
|
non_process
|
test local instrumented unit test for supervisedgrouplist activity children of issue testings for supervisedgrouplist and supervisedgrouppage supervisedgrouplist supervisedgrouppage
| 0
|
121,005
| 12,102,589,496
|
IssuesEvent
|
2020-04-20 16:56:40
|
minecode/website
|
https://api.github.com/repos/minecode/website
|
closed
|
Make release on actions
|
ci/cd documentation
|
Use https://github.com/actions/create-release (or like this) to make a new release when some pr is approved (integrate this step on github actions)
|
1.0
|
Make release on actions - Use https://github.com/actions/create-release (or like this) to make a new release when some pr is approved (integrate this step on github actions)
|
non_process
|
make release on actions use or like this to make a new release when some pr is approved integrate this step on github actions
| 0
|
694,671
| 23,824,341,108
|
IssuesEvent
|
2022-09-05 13:46:02
|
serverlessworkflow/synapse
|
https://api.github.com/repos/serverlessworkflow/synapse
|
closed
|
Edge is not drawn between default data case and transition to state
|
bug dashboard priority: medium weight: 1
|
**What happened**:
The edge is not drawn between default data case and transition to state
**What you expected to happen**:
Edge is drawn between the default data case and the state it should transition to
**How to reproduce it**:
Create a `switch` state which use a default data condition configuring a transition to another state
|
1.0
|
Edge is not drawn between default data case and transition to state - **What happened**:
The edge is not drawn between default data case and transition to state
**What you expected to happen**:
Edge is drawn between the default data case and the state it should transition to
**How to reproduce it**:
Create a `switch` state which use a default data condition configuring a transition to another state
|
non_process
|
edge is not drawn between default data case and transition to state what happened the edge is not drawn between default data case and transition to state what you expected to happen edge is drawn between the default data case and the state it should transition to how to reproduce it create a switch state which use a default data condition configuring a transition to another state
| 0
|
48,059
| 7,373,031,078
|
IssuesEvent
|
2018-03-13 16:11:30
|
umbraco/UmbracoDocs
|
https://api.github.com/repos/umbraco/UmbracoDocs
|
closed
|
Cloud Docs: Documentation about adding/removing environments
|
Umbraco Cloud missing documentation
|
Missing documentation about how to add / remove environment on a Cloud project.
Should be added either here:
https://our.umbraco.org/documentation/Umbraco-Cloud/Getting-Started/The-Umbraco-Cloud-Portal/
or here:
https://our.umbraco.org/documentation/Umbraco-Cloud/Getting-Started/Project-Overview/
|
1.0
|
Cloud Docs: Documentation about adding/removing environments - Missing documentation about how to add / remove environment on a Cloud project.
Should be added either here:
https://our.umbraco.org/documentation/Umbraco-Cloud/Getting-Started/The-Umbraco-Cloud-Portal/
or here:
https://our.umbraco.org/documentation/Umbraco-Cloud/Getting-Started/Project-Overview/
|
non_process
|
cloud docs documentation about adding removing environments missing documentation about how to add remove environment on a cloud project should be added either here or here
| 0
|
16,096
| 20,267,847,879
|
IssuesEvent
|
2022-02-15 13:45:23
|
sysflow-telemetry/sysflow
|
https://api.github.com/repos/sysflow-telemetry/sysflow
|
closed
|
Clarify and streamline engine modes and rule actions in sf-processor
|
enhancement sf-processor
|
Project: sf-processor
Sf-processor provides two options to control rule-driven filtering, record tagging and alert generation. The first parameter is the policy engine mode that is set in the processor pipeline specification. The second parameter is the rule action that can be set individually for each each rule.
Currently it is not straight forward and obvious what mode/actions exactly do and in what combination they can be used. Additionally, some current settings, like the `alert` action tag are not actively doing anything.
Desired is an approach that is ideally more intuitive and streamlined without redundancies.
The past and proposed functionality is tracked in https://github.com/sysflow-telemetry/sysflow/wiki/Processor-Modes-and-Rule-Actions
Discussion around this can take place in this issue.
|
1.0
|
Clarify and streamline engine modes and rule actions in sf-processor - Project: sf-processor
Sf-processor provides two options to control rule-driven filtering, record tagging and alert generation. The first parameter is the policy engine mode that is set in the processor pipeline specification. The second parameter is the rule action that can be set individually for each each rule.
Currently it is not straight forward and obvious what mode/actions exactly do and in what combination they can be used. Additionally, some current settings, like the `alert` action tag are not actively doing anything.
Desired is an approach that is ideally more intuitive and streamlined without redundancies.
The past and proposed functionality is tracked in https://github.com/sysflow-telemetry/sysflow/wiki/Processor-Modes-and-Rule-Actions
Discussion around this can take place in this issue.
|
process
|
clarify and streamline engine modes and rule actions in sf processor project sf processor sf processor provides two options to control rule driven filtering record tagging and alert generation the first parameter is the policy engine mode that is set in the processor pipeline specification the second parameter is the rule action that can be set individually for each each rule currently it is not straight forward and obvious what mode actions exactly do and in what combination they can be used additionally some current settings like the alert action tag are not actively doing anything desired is an approach that is ideally more intuitive and streamlined without redundancies the past and proposed functionality is tracked in discussion around this can take place in this issue
| 1
|
6,546
| 9,635,259,749
|
IssuesEvent
|
2019-05-16 00:17:41
|
googleapis/release-please
|
https://api.github.com/repos/googleapis/release-please
|
closed
|
chore(release): proposal for next release
|
release-candidate type: process
|
_:robot: Here's what the next release of **release-please** would look like._
---
### [1.3.1](https://www.github.com/googleapis/release-please/compare/v1.3.0...v1.3.1) (2019-05-16)
### Bug Fixes
* GitHub issues do not allow 'link comments' ([#92](https://www.github.com/googleapis/release-please/issues/92)) ([fe4cd4f](https://www.github.com/googleapis/release-please/commit/fe4cd4f))
* Node 10 is required for async/await ([#89](https://www.github.com/googleapis/release-please/issues/89)) ([c795eef](https://www.github.com/googleapis/release-please/commit/c795eef))
---
[//]: # footer follows.
* [ ] **Should I create this release for you :robot:?**
|
1.0
|
chore(release): proposal for next release - _:robot: Here's what the next release of **release-please** would look like._
---
### [1.3.1](https://www.github.com/googleapis/release-please/compare/v1.3.0...v1.3.1) (2019-05-16)
### Bug Fixes
* GitHub issues do not allow 'link comments' ([#92](https://www.github.com/googleapis/release-please/issues/92)) ([fe4cd4f](https://www.github.com/googleapis/release-please/commit/fe4cd4f))
* Node 10 is required for async/await ([#89](https://www.github.com/googleapis/release-please/issues/89)) ([c795eef](https://www.github.com/googleapis/release-please/commit/c795eef))
---
[//]: # footer follows.
* [ ] **Should I create this release for you :robot:?**
|
process
|
chore release proposal for next release robot here s what the next release of release please would look like bug fixes github issues do not allow link comments node is required for async await footer follows should i create this release for you robot
| 1
|
17,262
| 23,043,516,720
|
IssuesEvent
|
2022-07-23 14:27:35
|
andrewzah/openbook
|
https://api.github.com/repos/andrewzah/openbook
|
opened
|
get QR codes working
|
investigate rust-preprocessor lilypond
|
* [ ] figure out how to consistently overlay QR codes
* [ ] support multiple qr codes (1-3)
* [ ] create script to generate QR codes automatically
* [ ] pull links from song metadata, not hardcoded values
|
1.0
|
get QR codes working - * [ ] figure out how to consistently overlay QR codes
* [ ] support multiple qr codes (1-3)
* [ ] create script to generate QR codes automatically
* [ ] pull links from song metadata, not hardcoded values
|
process
|
get qr codes working figure out how to consistently overlay qr codes support multiple qr codes create script to generate qr codes automatically pull links from song metadata not hardcoded values
| 1
|
15,215
| 19,061,847,365
|
IssuesEvent
|
2021-11-26 08:53:37
|
metallb/metallb
|
https://api.github.com/repos/metallb/metallb
|
closed
|
Add Federico to CNCF maintainers list
|
process
|
We just added @fedepaol to the maintainers list for MetalLB. We need to add him on the CNCF side, as well. @rata volunteered to handle this.
|
1.0
|
Add Federico to CNCF maintainers list - We just added @fedepaol to the maintainers list for MetalLB. We need to add him on the CNCF side, as well. @rata volunteered to handle this.
|
process
|
add federico to cncf maintainers list we just added fedepaol to the maintainers list for metallb we need to add him on the cncf side as well rata volunteered to handle this
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.