Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
741,724
25,814,585,605
IssuesEvent
2022-12-12 03:10:59
containerd/nerdctl
https://api.github.com/repos/containerd/nerdctl
closed
[Rootful + SELinux] [BuildKit] RUN command throws error code 139 on centos
bug priority/high kind/external
### Description Cannot build using containerfile with RUN command on centos. The same containerfile works fine on ubuntu. ### Steps to reproduce the issue 1. Have a simple Containter file: FROM alpine RUN ls 2. Build it ### Describe the results you received and expected nerdctl build -t test:1 . [+] Building 1.5s (4/5) [+] Building 1.6s (5/5) FINISHED => [internal] load build definition from Containerfile 0.0s => => transferring dockerfile: 118B 0.0s => [internal] load .dockerignore 0.0s => => transferring context: 2B 0.0s => [internal] load metadata for docker.io/library/alpine:latest 1.0s => [1/2] FROM docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.3s => => resolve docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.0s => => sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 2.81MB / 2.81MB 0.1s => => extracting sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 0.1s => ERROR [2/2] RUN ls 0.2s ------ > [2/2] RUN ls: ------ Containerfile:2 -------------------- 1 | FROM alpine 2 | >>> RUN ls 3 | -------------------- error: failed to solve: process "/bin/sh -c ls" did not complete successfully: exit code: 139 FATA[0001] unrecognized image format ### What version of nerdctl are you using? nerdctl version 0.23.0 ### Are you using a variant of nerdctl? (e.g., Rancher Desktop) _No response_ ### Host information Client: Namespace: default Debug Mode: false Server: Server Version: v1.6.8 Storage Driver: overlayfs Logging Driver: json-file Cgroup Driver: cgroupfs Cgroup Version: 1 Plugins: Log: fluentd journald json-file Storage: native overlayfs Security Options: seccomp Profile: default Kernel Version: 4.18.0-408.el8.x86_64 Operating System: CentOS Stream 8 OSType: linux Architecture: x86_64 CPUs: 2 Total Memory: 14.47GiB Name: pcr-nerdctl-tst2 ID: 2908f464-ed9f-4476-99cc-205a294e6e51 WARNING: bridge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled
1.0
[Rootful + SELinux] [BuildKit] RUN command throws error code 139 on centos - ### Description Cannot build using containerfile with RUN command on centos. The same containerfile works fine on ubuntu. ### Steps to reproduce the issue 1. Have a simple Containter file: FROM alpine RUN ls 2. Build it ### Describe the results you received and expected nerdctl build -t test:1 . [+] Building 1.5s (4/5) [+] Building 1.6s (5/5) FINISHED => [internal] load build definition from Containerfile 0.0s => => transferring dockerfile: 118B 0.0s => [internal] load .dockerignore 0.0s => => transferring context: 2B 0.0s => [internal] load metadata for docker.io/library/alpine:latest 1.0s => [1/2] FROM docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.3s => => resolve docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.0s => => sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 2.81MB / 2.81MB 0.1s => => extracting sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 0.1s => ERROR [2/2] RUN ls 0.2s ------ > [2/2] RUN ls: ------ Containerfile:2 -------------------- 1 | FROM alpine 2 | >>> RUN ls 3 | -------------------- error: failed to solve: process "/bin/sh -c ls" did not complete successfully: exit code: 139 FATA[0001] unrecognized image format ### What version of nerdctl are you using? nerdctl version 0.23.0 ### Are you using a variant of nerdctl? (e.g., Rancher Desktop) _No response_ ### Host information Client: Namespace: default Debug Mode: false Server: Server Version: v1.6.8 Storage Driver: overlayfs Logging Driver: json-file Cgroup Driver: cgroupfs Cgroup Version: 1 Plugins: Log: fluentd journald json-file Storage: native overlayfs Security Options: seccomp Profile: default Kernel Version: 4.18.0-408.el8.x86_64 Operating System: CentOS Stream 8 OSType: linux Architecture: x86_64 CPUs: 2 Total Memory: 14.47GiB Name: pcr-nerdctl-tst2 ID: 2908f464-ed9f-4476-99cc-205a294e6e51 WARNING: bridge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled
priority
run command throws error code on centos description cannot build using containerfile with run command on centos the same containerfile works fine on ubuntu steps to reproduce the issue have a simple containter file from alpine run ls build it describe the results you received and expected nerdctl build t test building building finished load build definition from containerfile transferring dockerfile load dockerignore transferring context load metadata for docker io library alpine latest from docker io library alpine resolve docker io library alpine extracting error run ls run ls containerfile from alpine run ls error failed to solve process bin sh c ls did not complete successfully exit code fata unrecognized image format what version of nerdctl are you using nerdctl version are you using a variant of nerdctl e g rancher desktop no response host information client namespace default debug mode false server server version storage driver overlayfs logging driver json file cgroup driver cgroupfs cgroup version plugins log fluentd journald json file storage native overlayfs security options seccomp profile default kernel version operating system centos stream ostype linux architecture cpus total memory name pcr nerdctl id warning bridge nf call iptables is disabled warning bridge nf call is disabled
1
566,104
16,796,077,732
IssuesEvent
2021-06-16 03:52:40
sodafoundation/multi-cloud
https://api.github.com/repos/sodafoundation/multi-cloud
opened
AccessKey , SecretKey to be created from API
High Priority
**Issue/Feature Description:** AccessKey , SecretKey to be created from API - To enable users to create AKSK from the API , rather than the UI. AkSk from API will enable users to create the AKSK from the API and integrate to their Clients. **Why this issue to fixed / feature is needed(give scenarios or use cases):** AkSk from API will enable users to create the AKSK from the API and integrate to their Clients. **How to reproduce, in case of a bug:** N/A **Other Notes / Environment Information: (Please give the env information, log link or any useful information for this issue)**
1.0
AccessKey , SecretKey to be created from API - **Issue/Feature Description:** AccessKey , SecretKey to be created from API - To enable users to create AKSK from the API , rather than the UI. AkSk from API will enable users to create the AKSK from the API and integrate to their Clients. **Why this issue to fixed / feature is needed(give scenarios or use cases):** AkSk from API will enable users to create the AKSK from the API and integrate to their Clients. **How to reproduce, in case of a bug:** N/A **Other Notes / Environment Information: (Please give the env information, log link or any useful information for this issue)**
priority
accesskey secretkey to be created from api issue feature description accesskey secretkey to be created from api to enable users to create aksk from the api rather than the ui aksk from api will enable users to create the aksk from the api and integrate to their clients why this issue to fixed feature is needed give scenarios or use cases aksk from api will enable users to create the aksk from the api and integrate to their clients how to reproduce in case of a bug n a other notes environment information please give the env information log link or any useful information for this issue
1
96,091
3,964,411,317
IssuesEvent
2016-05-03 00:45:32
daronco/test-issue-migrate2
https://api.github.com/repos/daronco/test-issue-migrate2
closed
Set a numeric voiceBridge for every room
Priority: High Status: Resolved Type: Bug
--- Author Name: **Leonardo Daronco** (@daronco) Original Redmine Issue: 145, http://dev.mconf.org/redmine/issues/145 Original Assignee: Leonardo Daronco --- @Google user: leonardo...@gmail.com@ Use a numeric "voiceBridge" param when creating a room. From the BigBlueButton API documentation: "we recommend you always pass a 5 digit voiceBridge parameter -- and have it begin with the digit '7' if you are using the default FreeSWITCH setup" So the voiceBridge should have 5 digits, the first one being a '7' and the last ones can be taken from the meetingID (after implementing "/p/mconf/issues/detail?id=38": issue #38 ) or can be random (but unique).
1.0
Set a numeric voiceBridge for every room - --- Author Name: **Leonardo Daronco** (@daronco) Original Redmine Issue: 145, http://dev.mconf.org/redmine/issues/145 Original Assignee: Leonardo Daronco --- @Google user: leonardo...@gmail.com@ Use a numeric "voiceBridge" param when creating a room. From the BigBlueButton API documentation: "we recommend you always pass a 5 digit voiceBridge parameter -- and have it begin with the digit '7' if you are using the default FreeSWITCH setup" So the voiceBridge should have 5 digits, the first one being a '7' and the last ones can be taken from the meetingID (after implementing "/p/mconf/issues/detail?id=38": issue #38 ) or can be random (but unique).
priority
set a numeric voicebridge for every room author name leonardo daronco daronco original redmine issue original assignee leonardo daronco google user leonardo gmail com use a numeric voicebridge param when creating a room from the bigbluebutton api documentation we recommend you always pass a digit voicebridge parameter and have it begin with the digit if you are using the default freeswitch setup so the voicebridge should have digits the first one being a and the last ones can be taken from the meetingid after implementing p mconf issues detail id issue or can be random but unique
1
425,104
12,335,783,449
IssuesEvent
2020-05-14 12:34:46
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
opened
Possible regression in job activation
Impact: Performance Scope: broker Severity: High Status: Needs Priority Type: Bug
**Describe the bug** I observed a big drop in throughput on running our normal benchmark. ![general](https://user-images.githubusercontent.com/2758593/81934619-8b005a80-95ef-11ea-82d8-2dfd4b2a5f50.png) I would normally expect ~200 workflows and task to be completed. It seems that job activation is the problem since the activation latency is quite high. ![activation](https://user-images.githubusercontent.com/2758593/81934753-c1d67080-95ef-11ea-9218-d4fc8972e175.png) The standalone gateway also throws endless the following timeouts: ``` 2020-05-14 12:31:23.930 [io.zeebe.gateway.impl.broker.BrokerRequestManager] [gateway-scheduler-zb-actors-0] ERROR io.zeebe.gateway - Error handling gRPC request io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: Time out between gateway and broker: Request type command-api-1 timed out in 15000 milliseconds at io.grpc.Status.asRuntimeException(Status.java:524) ~[grpc-api-1.29.0.jar:1.29.0] at io.zeebe.gateway.EndpointManager.convertThrowable(EndpointManager.java:397) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.EndpointManager.lambda$sendRequest$3(EndpointManager.java:311) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequest$3(BrokerRequestManager.java:148) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequestInternal$5(BrokerRequestManager.java:191) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.future.FutureContinuationRunnable.run(FutureContinuationRunnable.java:33) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorJob.invoke(ActorJob.java:76) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorJob.execute(ActorJob.java:39) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorTask.execute(ActorTask.java:118) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:107) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.doWork(ActorThread.java:91) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.run(ActorThread.java:204) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] Caused by: java.util.concurrent.TimeoutException: Request type command-api-1 timed out in 15000 milliseconds at io.atomix.cluster.messaging.impl.AbstractClientConnection$Callback.timeout(AbstractClientConnection.java:163) ~[atomix-cluster-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) ~[?:?] at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown Source) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) ~[?:?] at java.lang.Thread.run(Unknown Source) ~[?:?] ``` **To Reproduce** Run the helm chart (v100) with our benchmark, **Expected behavior** Around 200 workflows are completed per second.
1.0
Possible regression in job activation - **Describe the bug** I observed a big drop in throughput on running our normal benchmark. ![general](https://user-images.githubusercontent.com/2758593/81934619-8b005a80-95ef-11ea-82d8-2dfd4b2a5f50.png) I would normally expect ~200 workflows and task to be completed. It seems that job activation is the problem since the activation latency is quite high. ![activation](https://user-images.githubusercontent.com/2758593/81934753-c1d67080-95ef-11ea-9218-d4fc8972e175.png) The standalone gateway also throws endless the following timeouts: ``` 2020-05-14 12:31:23.930 [io.zeebe.gateway.impl.broker.BrokerRequestManager] [gateway-scheduler-zb-actors-0] ERROR io.zeebe.gateway - Error handling gRPC request io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: Time out between gateway and broker: Request type command-api-1 timed out in 15000 milliseconds at io.grpc.Status.asRuntimeException(Status.java:524) ~[grpc-api-1.29.0.jar:1.29.0] at io.zeebe.gateway.EndpointManager.convertThrowable(EndpointManager.java:397) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.EndpointManager.lambda$sendRequest$3(EndpointManager.java:311) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequest$3(BrokerRequestManager.java:148) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequestInternal$5(BrokerRequestManager.java:191) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.future.FutureContinuationRunnable.run(FutureContinuationRunnable.java:33) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorJob.invoke(ActorJob.java:76) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorJob.execute(ActorJob.java:39) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorTask.execute(ActorTask.java:118) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:107) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.doWork(ActorThread.java:91) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at io.zeebe.util.sched.ActorThread.run(ActorThread.java:204) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] Caused by: java.util.concurrent.TimeoutException: Request type command-api-1 timed out in 15000 milliseconds at io.atomix.cluster.messaging.impl.AbstractClientConnection$Callback.timeout(AbstractClientConnection.java:163) ~[atomix-cluster-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) ~[?:?] at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown Source) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) ~[?:?] at java.lang.Thread.run(Unknown Source) ~[?:?] ``` **To Reproduce** Run the helm chart (v100) with our benchmark, **Expected behavior** Around 200 workflows are completed per second.
priority
possible regression in job activation describe the bug i observed a big drop in throughput on running our normal benchmark i would normally expect workflows and task to be completed it seems that job activation is the problem since the activation latency is quite high the standalone gateway also throws endless the following timeouts error io zeebe gateway error handling grpc request io grpc statusruntimeexception deadline exceeded time out between gateway and broker request type command api timed out in milliseconds at io grpc status asruntimeexception status java at io zeebe gateway endpointmanager convertthrowable endpointmanager java at io zeebe gateway endpointmanager lambda sendrequest endpointmanager java at io zeebe gateway impl broker brokerrequestmanager lambda sendrequest brokerrequestmanager java at io zeebe gateway impl broker brokerrequestmanager lambda sendrequestinternal brokerrequestmanager java at io zeebe util sched future futurecontinuationrunnable run futurecontinuationrunnable java at io zeebe util sched actorjob invoke actorjob java at io zeebe util sched actorjob execute actorjob java at io zeebe util sched actortask execute actortask java at io zeebe util sched actorthread executecurrenttask actorthread java at io zeebe util sched actorthread dowork actorthread java at io zeebe util sched actorthread run actorthread java caused by java util concurrent timeoutexception request type command api timed out in milliseconds at io atomix cluster messaging impl abstractclientconnection callback timeout abstractclientconnection java at java util concurrent executors runnableadapter call unknown source at java util concurrent futuretask run unknown source at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run unknown source at java util concurrent threadpoolexecutor runworker unknown source at java util concurrent threadpoolexecutor worker run unknown source at java lang thread run unknown source to reproduce run the helm chart with our benchmark expected behavior around workflows are completed per second
1
619,804
19,535,416,577
IssuesEvent
2021-12-31 05:08:00
ajayyy/SponsorBlock
https://api.github.com/repos/ajayyy/SponsorBlock
closed
Better-sqlite3 and other improvements
HIGH PRIORITY
[better-sqlite](https://github.com/JoshuaWise/better-sqlite3) is faster than the more popular `sqlite3` node library especially when load increases. The reason is that sqlite3 is "asynchronous" even though all the computation still happens on the same thread. This means the only thing it being async does is it increases the overhead significantly. (The reason it's more popular is as far as I know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes.)
1.0
Better-sqlite3 and other improvements - [better-sqlite](https://github.com/JoshuaWise/better-sqlite3) is faster than the more popular `sqlite3` node library especially when load increases. The reason is that sqlite3 is "asynchronous" even though all the computation still happens on the same thread. This means the only thing it being async does is it increases the overhead significantly. (The reason it's more popular is as far as I know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes.)
priority
better and other improvements is faster than the more popular node library especially when load increases the reason is that is asynchronous even though all the computation still happens on the same thread this means the only thing it being async does is it increases the overhead significantly the reason it s more popular is as far as i know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes
1
523,524
15,184,319,187
IssuesEvent
2021-02-15 09:24:54
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Formatter] Formatting doesn't work properly when a part of the string is in a different line
Area/Formatting Priority/High Team/Tooling Type/Bug
**Description:** Please consider the following scenario ![Screenshot from 2021-01-22 12-22-52](https://user-images.githubusercontent.com/19838787/105457309-993d6500-5cac-11eb-9c91-f9cbaa647e39.png) When I format the above code snippet additional spaces are added to the string values as given below. ![Screenshot from 2021-01-22 12-24-06](https://user-images.githubusercontent.com/19838787/105457424-d0ac1180-5cac-11eb-8011-3053fafe45da.png) **Steps to reproduce:** **Affected Versions:** SLP9-snapshot **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
[Formatter] Formatting doesn't work properly when a part of the string is in a different line - **Description:** Please consider the following scenario ![Screenshot from 2021-01-22 12-22-52](https://user-images.githubusercontent.com/19838787/105457309-993d6500-5cac-11eb-9c91-f9cbaa647e39.png) When I format the above code snippet additional spaces are added to the string values as given below. ![Screenshot from 2021-01-22 12-24-06](https://user-images.githubusercontent.com/19838787/105457424-d0ac1180-5cac-11eb-8011-3053fafe45da.png) **Steps to reproduce:** **Affected Versions:** SLP9-snapshot **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
priority
formatting doesn t work properly when a part of the string is in a different line description please consider the following scenario when i format the above code snippet additional spaces are added to the string values as given below steps to reproduce affected versions snapshot os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
1
393,662
11,623,112,634
IssuesEvent
2020-02-27 08:15:58
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Map water doesnt not always load on web interface
Priority: High Status: Fixed
Repeatedly reload any page with a map, sometimes the water does not load correctly. https://gyazo.com/ce4aae1286389ebf4692477e4c662d25
1.0
Map water doesnt not always load on web interface - Repeatedly reload any page with a map, sometimes the water does not load correctly. https://gyazo.com/ce4aae1286389ebf4692477e4c662d25
priority
map water doesnt not always load on web interface repeatedly reload any page with a map sometimes the water does not load correctly
1
65,004
3,222,593,846
IssuesEvent
2015-10-09 02:33:50
cs2103aug2015-t11-4j/main
https://api.github.com/repos/cs2103aug2015-t11-4j/main
closed
Parser support for update task
:parser priority.high type.task
Takes in string value from Logic and edit the certain task object and return the object to Logic
1.0
Parser support for update task - Takes in string value from Logic and edit the certain task object and return the object to Logic
priority
parser support for update task takes in string value from logic and edit the certain task object and return the object to logic
1
398,475
11,741,500,717
IssuesEvent
2020-03-11 21:54:13
SacredDuckwhale/Rarity
https://api.github.com/repos/SacredDuckwhale/Rarity
opened
The Combat-log based attempts detection doesn't work properly when used to detect outdoor world bosses
complexity: moderate module:core priority:high status:accepted type:bug
Verified for Dunegorger Kraulok. After looking at the code, it appears any UNIT_DIED event is now triggering the detection, when only those that are caused by the player or their party/raid should count. The implementation appears to rely on bit flags for a certain srcFlag bitmap field (see https://wow.gamepedia.com/API_CombatLogGetCurrentEventInfo), but those are no longer working? The value is always -2147483648 and so ALL kills count regardless of who actually caused them. Potential solutions: * Repair the bit flag detection, if Blizzard hasn't removed/broken it * Failing that, it might be possible to rely on the defeat detection (loot lockout) and count an attempt if and only if the player isn't yet logged out Notes: * MOP world bosses are probably affected, as well
1.0
The Combat-log based attempts detection doesn't work properly when used to detect outdoor world bosses - Verified for Dunegorger Kraulok. After looking at the code, it appears any UNIT_DIED event is now triggering the detection, when only those that are caused by the player or their party/raid should count. The implementation appears to rely on bit flags for a certain srcFlag bitmap field (see https://wow.gamepedia.com/API_CombatLogGetCurrentEventInfo), but those are no longer working? The value is always -2147483648 and so ALL kills count regardless of who actually caused them. Potential solutions: * Repair the bit flag detection, if Blizzard hasn't removed/broken it * Failing that, it might be possible to rely on the defeat detection (loot lockout) and count an attempt if and only if the player isn't yet logged out Notes: * MOP world bosses are probably affected, as well
priority
the combat log based attempts detection doesn t work properly when used to detect outdoor world bosses verified for dunegorger kraulok after looking at the code it appears any unit died event is now triggering the detection when only those that are caused by the player or their party raid should count the implementation appears to rely on bit flags for a certain srcflag bitmap field see but those are no longer working the value is always and so all kills count regardless of who actually caused them potential solutions repair the bit flag detection if blizzard hasn t removed broken it failing that it might be possible to rely on the defeat detection loot lockout and count an attempt if and only if the player isn t yet logged out notes mop world bosses are probably affected as well
1
810,374
30,239,070,418
IssuesEvent
2023-07-06 12:22:40
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Cypress image snapshots not working
Bug :lady_beetle: Sprint Priority: High Frontend :sunglasses:
**Describe the bug** Cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the UI. **To Reproduce** Steps to reproduce the behavior: - In the new translations UI, in the component for the translations lists, change the type for the action buttons, so that they visually change. - Run the E2E relevant to translations. - There’s no error for the visual change. **Expected behavior** We should have a very low threshold of tolerance for visual changes in the UI and the test should fail.
1.0
Cypress image snapshots not working - **Describe the bug** Cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the UI. **To Reproduce** Steps to reproduce the behavior: - In the new translations UI, in the component for the translations lists, change the type for the action buttons, so that they visually change. - Run the E2E relevant to translations. - There’s no error for the visual change. **Expected behavior** We should have a very low threshold of tolerance for visual changes in the UI and the test should fail.
priority
cypress image snapshots not working describe the bug cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the ui to reproduce steps to reproduce the behavior in the new translations ui in the component for the translations lists change the type for the action buttons so that they visually change run the relevant to translations there’s no error for the visual change expected behavior we should have a very low threshold of tolerance for visual changes in the ui and the test should fail
1
423,632
12,299,364,335
IssuesEvent
2020-05-11 12:15:21
bounswe/bounswe2020group1
https://api.github.com/repos/bounswe/bounswe2020group1
closed
Implement the functions for "find similar words" API
priority:high type:implementation
Implementing functions that use the API provided by Datamuse.
1.0
Implement the functions for "find similar words" API - Implementing functions that use the API provided by Datamuse.
priority
implement the functions for find similar words api implementing functions that use the api provided by datamuse
1
808,195
30,037,669,465
IssuesEvent
2023-06-27 13:43:39
tum-esm/hermes
https://api.github.com/repos/tum-esm/hermes
closed
Send in-flow sensor measurement with every CO2 measurement over MQTT
type:feature status:implemented high-priority scope:sensor
The frequency of the in-flow measurements needs to be higher. (Current implementation every 2 minutes) Change to directly integrate the in-flow sensor data into the CO2 measurement MQTT data stream. This also makes it more visible what sensor data the CO2 sensor received to perform the correction and allows us to do it ourselves in the future. Physical Background: The pump creates a flow between 0.4-0.6 ppm. Depending on the flow the underpressure in the system changes. This change has a direct influence on the correction based on pressure in the Vaisala GMP343.
1.0
Send in-flow sensor measurement with every CO2 measurement over MQTT - The frequency of the in-flow measurements needs to be higher. (Current implementation every 2 minutes) Change to directly integrate the in-flow sensor data into the CO2 measurement MQTT data stream. This also makes it more visible what sensor data the CO2 sensor received to perform the correction and allows us to do it ourselves in the future. Physical Background: The pump creates a flow between 0.4-0.6 ppm. Depending on the flow the underpressure in the system changes. This change has a direct influence on the correction based on pressure in the Vaisala GMP343.
priority
send in flow sensor measurement with every measurement over mqtt the frequency of the in flow measurements needs to be higher current implementation every minutes change to directly integrate the in flow sensor data into the measurement mqtt data stream this also makes it more visible what sensor data the sensor received to perform the correction and allows us to do it ourselves in the future physical background the pump creates a flow between ppm depending on the flow the underpressure in the system changes this change has a direct influence on the correction based on pressure in the vaisala
1
792,351
27,956,882,081
IssuesEvent
2023-03-24 13:03:33
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Needs to get the default value of an input parameter in a resource function
Type/Improvement Priority/High Team/jBallerina
**Description:** When there is an input parameter with a default value in a resource function, there should be a way to get the default value of this parameter. This is a requirement for the Ballerina GraphQL package implementation. Related: #27417, https://github.com/ballerina-platform/ballerina-standard-library/issues/1266
1.0
Needs to get the default value of an input parameter in a resource function - **Description:** When there is an input parameter with a default value in a resource function, there should be a way to get the default value of this parameter. This is a requirement for the Ballerina GraphQL package implementation. Related: #27417, https://github.com/ballerina-platform/ballerina-standard-library/issues/1266
priority
needs to get the default value of an input parameter in a resource function description when there is an input parameter with a default value in a resource function there should be a way to get the default value of this parameter this is a requirement for the ballerina graphql package implementation related
1
484,359
13,938,471,195
IssuesEvent
2020-10-22 15:18:24
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
Pattern based anomaly detection engine for the cloud
Complexity/High Component/Analytics Priority/Low gateway research
**Is your feature request related to a problem? Please describe.** $subject allows monitoring an Identity Server cloud deployment for unusual behaviors, based on pattern-based anomalies of various signals coming from the deployment. The engine can be used to trigger alerts in near-realtime, on potential malicious requests, attacks, and other unexpected activities. **Describe the solution you would prefer** Come up with an anomaly detection engine PoC, where various signals from the Identity Server deployment(request rate, authentication events, etc) can be analyzed in general, and compare it with the patterns observed in the past, to identify anomalies. The engine should support various un-supervised analysis mechanisms, due to the vast amount of data that can be generated from a cloud deployment. **Additional context** -
1.0
Pattern based anomaly detection engine for the cloud - **Is your feature request related to a problem? Please describe.** $subject allows monitoring an Identity Server cloud deployment for unusual behaviors, based on pattern-based anomalies of various signals coming from the deployment. The engine can be used to trigger alerts in near-realtime, on potential malicious requests, attacks, and other unexpected activities. **Describe the solution you would prefer** Come up with an anomaly detection engine PoC, where various signals from the Identity Server deployment(request rate, authentication events, etc) can be analyzed in general, and compare it with the patterns observed in the past, to identify anomalies. The engine should support various un-supervised analysis mechanisms, due to the vast amount of data that can be generated from a cloud deployment. **Additional context** -
priority
pattern based anomaly detection engine for the cloud is your feature request related to a problem please describe subject allows monitoring an identity server cloud deployment for unusual behaviors based on pattern based anomalies of various signals coming from the deployment the engine can be used to trigger alerts in near realtime on potential malicious requests attacks and other unexpected activities describe the solution you would prefer come up with an anomaly detection engine poc where various signals from the identity server deployment request rate authentication events etc can be analyzed in general and compare it with the patterns observed in the past to identify anomalies the engine should support various un supervised analysis mechanisms due to the vast amount of data that can be generated from a cloud deployment additional context
1
184,914
6,717,386,738
IssuesEvent
2017-10-14 20:35:20
semperfiwebdesign/all-in-one-seo-pack
https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack
opened
Exclude Pages option no longer working
Initial Review Priority | High
As reported by Albert Belzer (belzer9@gmail.com) on October 14, 2017. In his original e-mail, Albert states that both the "Disable SEO for this post/page option" setting on the Edit screen and the "Exlude Pages" setting in the General Settings are not working. I checked this and although the former is working, the latter seems like it's not.
1.0
Exclude Pages option no longer working - As reported by Albert Belzer (belzer9@gmail.com) on October 14, 2017. In his original e-mail, Albert states that both the "Disable SEO for this post/page option" setting on the Edit screen and the "Exlude Pages" setting in the General Settings are not working. I checked this and although the former is working, the latter seems like it's not.
priority
exclude pages option no longer working as reported by albert belzer gmail com on october in his original e mail albert states that both the disable seo for this post page option setting on the edit screen and the exlude pages setting in the general settings are not working i checked this and although the former is working the latter seems like it s not
1
382,199
11,302,244,285
IssuesEvent
2020-01-17 17:12:56
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
reopened
Prototype APIs can't be used for API products
3.1.0 Priority/Highest Resolution/Fixed Type/Bug
$subject. The inline script is not copied to the API product synapse. Hense invocation fails.
1.0
Prototype APIs can't be used for API products - $subject. The inline script is not copied to the API product synapse. Hense invocation fails.
priority
prototype apis can t be used for api products subject the inline script is not copied to the api product synapse hense invocation fails
1
382,645
11,309,760,056
IssuesEvent
2020-01-19 15:16:11
FederatedAI/FATE
https://api.github.com/repos/FederatedAI/FATE
closed
Support for secret sharing scheme
enhancement priority:high research
**Is your feature request related to a problem? Please describe.** Secret sharing Scheme is a must-have for FATE project. **Describe the solution you'd like** Do R&D on implementing secret sharing operations, such as: 1. Create beaver triple 2. Add, Multiply, Division, Compare, and others Having secret sharing operations been created, then 1. Implement secret sharing based LR 2. Implement secret sharing based FTL These works do not need to be full-fledged for industrial applications. However, they should be able to help us create various secure federated learning algorithms/prototypes.
1.0
Support for secret sharing scheme - **Is your feature request related to a problem? Please describe.** Secret sharing Scheme is a must-have for FATE project. **Describe the solution you'd like** Do R&D on implementing secret sharing operations, such as: 1. Create beaver triple 2. Add, Multiply, Division, Compare, and others Having secret sharing operations been created, then 1. Implement secret sharing based LR 2. Implement secret sharing based FTL These works do not need to be full-fledged for industrial applications. However, they should be able to help us create various secure federated learning algorithms/prototypes.
priority
support for secret sharing scheme is your feature request related to a problem please describe secret sharing scheme is a must have for fate project describe the solution you d like do r d on implementing secret sharing operations such as create beaver triple add multiply division compare and others having secret sharing operations been created then implement secret sharing based lr implement secret sharing based ftl these works do not need to be full fledged for industrial applications however they should be able to help us create various secure federated learning algorithms prototypes
1
72,470
3,386,257,735
IssuesEvent
2015-11-27 16:22:22
CosmosOS/Cosmos
https://api.github.com/repos/CosmosOS/Cosmos
closed
Dup tries to pop more stuff from analytical stack than there is!
area_compiler complexity_medium pending_verification priority_high
Log: ``` 4> Error: Exception: System.Exception: Error compiling method 'SystemVoidKernelCommandsInputCommand': System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is! 4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369 4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714 4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 ---> System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is! 4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369 4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714 4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 4> --- End of inner exception stack trace --- 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 529 4> at Cosmos.IL2CPU.ILScanner.Assemble() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 944 4> at Cosmos.IL2CPU.ILScanner.Execute(MethodBase aStartMethod) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 256 4> at Cosmos.IL2CPU.CompilerEngine.Execute() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\CompilerEngine.cs:line 238 ``` And there is code where it heappen: ```C# public static void InputCommand() { Console.Write("D:/command>"); comd = Console.ReadLine(); comd = comd.ToLower(); if (comd == "reboot") h.Power.Restart(); else if (comd == "shutdown") h.Power.Shutdown(); else if (comd == "echo") { Console.Write("Echo>"); arg = Console.ReadLine(); Console.WriteLine(arg); } else if (comd == "notepad") { System.CLI.Applications.Notepad(); } else if (comd == "cls") { Console.Clear(); Console.WriteLine("TriangleOS"); Console.WriteLine("============================="); } else if (comd == "soundtest") { Console.Write("Frequency>"); arg = Console.ReadLine(); Console.Write("Duration>"); optarg = Console.ReadLine(); Console.Write("Eh, this isn't implemted right now."); //h.Multimedia.Speakers.CallSound(int.Parse(arg), int.Parse(optarg)); } else if (comd == "boot") { Console.WriteLine("Starting TriangleOS.Drivers . . ."); //ProcessManager.Process Audio = new ProcessManager.Process(); //ProcessManager.Process Graphics = new ProcessManager.Process(); //Graphics.ProcessThread = new System.Threading.Thread( h.Graphics.LowLevel.init(); //); //ProcessManager.Process Mouse = new ProcessManager.Process(); //Mouse.ProcessThread = new System.Threading.Thread( h.Mouse.InitMouse(); //); //Graphics.Start(); //Mouse.Start(); //Audio.ProcessThread = new System.Threading.Thread( h.Multimedia.Speakers.IntailizeAudio(); //); Kernel.GUI(); } else if (comd == "cliboot") { System.CLI.Controls.TextBox Text = new System.CLI.Controls.TextBox(); Text.x = 1; Text.y = 1; Text.length = 20; Text.DrawTextBox(); Text.TypeInto(); System.CLI.Controls.Button OK = new System.CLI.Controls.Button(); OK.y = 22; OK.x = 1; OK.width = 6; OK.height = 1; OK.text = "OK"; OK.DrawButton(); } else if (comd == "calculator") { System.CLI.Applications.Calculator(); } else if (comd == "cd") { h.Graphics.Console.ErrO("Impossible operation performed. Can't request I/O while it isn't running!"); } else if (comd == "dir") { Console.WriteLine("This isn't folder. you can use <cd> to go up folder."); } else if (comd == "paint") { System.CLI.Applications.Paint(); } else if (comd == "changelog") { Console.WriteLine("You are running version 0.0.3 Dev. Only Luka see the DEV!"); Console.WriteLine("v0.0.2:"); Console.WriteLine("Blue screen with cursor. not clearing."); Console.WriteLine("v0.0.3:"); Console.WriteLine("Command Line Shell with Broken CLI, but working unresponsive GUI, but with DOS-like Shell. I/O, Audio, Multithreading, Shutdown doesn't work."); } else if (comd == "help") { Console.WriteLine("Copyright 2015 Thontelix TriangleOS. Special thanks to Cosmos .net ASM Compiler."); Console.WriteLine("type CHANGELOG to get version changes."); Console.WriteLine("How to use:"); Console.WriteLine("After every typed command, press enter. the output of command will be detailed. If there is blinking bottom line cursor, then you need to input something, if it doesn't give any feedback, then its operating a activity. If you want GUI, type 'boot' and press enter."); Console.WriteLine("Commands:"); Console.WriteLine("shutdown - Gives you ability to safe turn off PC"); Console.WriteLine("boot - Boots you into TriangleOS"); Console.WriteLine("reboot - Reboots your PC"); Console.WriteLine("cd - Travel through directories"); Console.WriteLine("dir - Read content of directory"); Console.WriteLine("echo - Backs string you enter"); Console.WriteLine("shutdown - Backs string you enter"); Console.WriteLine("soundtest - Speakers Drivers. Doesnt work for now"); Console.WriteLine("help - Gives you list of commands."); Console.WriteLine("paint - Console Paint App. CAUTION:Not for epilepsy persons."); } else { Console.WriteLine("That command doesn't exist. :("); } } ```
1.0
Dup tries to pop more stuff from analytical stack than there is! - Log: ``` 4> Error: Exception: System.Exception: Error compiling method 'SystemVoidKernelCommandsInputCommand': System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is! 4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369 4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714 4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 ---> System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is! 4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369 4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714 4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 4> --- End of inner exception stack trace --- 4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 529 4> at Cosmos.IL2CPU.ILScanner.Assemble() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 944 4> at Cosmos.IL2CPU.ILScanner.Execute(MethodBase aStartMethod) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 256 4> at Cosmos.IL2CPU.CompilerEngine.Execute() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\CompilerEngine.cs:line 238 ``` And there is code where it heappen: ```C# public static void InputCommand() { Console.Write("D:/command>"); comd = Console.ReadLine(); comd = comd.ToLower(); if (comd == "reboot") h.Power.Restart(); else if (comd == "shutdown") h.Power.Shutdown(); else if (comd == "echo") { Console.Write("Echo>"); arg = Console.ReadLine(); Console.WriteLine(arg); } else if (comd == "notepad") { System.CLI.Applications.Notepad(); } else if (comd == "cls") { Console.Clear(); Console.WriteLine("TriangleOS"); Console.WriteLine("============================="); } else if (comd == "soundtest") { Console.Write("Frequency>"); arg = Console.ReadLine(); Console.Write("Duration>"); optarg = Console.ReadLine(); Console.Write("Eh, this isn't implemted right now."); //h.Multimedia.Speakers.CallSound(int.Parse(arg), int.Parse(optarg)); } else if (comd == "boot") { Console.WriteLine("Starting TriangleOS.Drivers . . ."); //ProcessManager.Process Audio = new ProcessManager.Process(); //ProcessManager.Process Graphics = new ProcessManager.Process(); //Graphics.ProcessThread = new System.Threading.Thread( h.Graphics.LowLevel.init(); //); //ProcessManager.Process Mouse = new ProcessManager.Process(); //Mouse.ProcessThread = new System.Threading.Thread( h.Mouse.InitMouse(); //); //Graphics.Start(); //Mouse.Start(); //Audio.ProcessThread = new System.Threading.Thread( h.Multimedia.Speakers.IntailizeAudio(); //); Kernel.GUI(); } else if (comd == "cliboot") { System.CLI.Controls.TextBox Text = new System.CLI.Controls.TextBox(); Text.x = 1; Text.y = 1; Text.length = 20; Text.DrawTextBox(); Text.TypeInto(); System.CLI.Controls.Button OK = new System.CLI.Controls.Button(); OK.y = 22; OK.x = 1; OK.width = 6; OK.height = 1; OK.text = "OK"; OK.DrawButton(); } else if (comd == "calculator") { System.CLI.Applications.Calculator(); } else if (comd == "cd") { h.Graphics.Console.ErrO("Impossible operation performed. Can't request I/O while it isn't running!"); } else if (comd == "dir") { Console.WriteLine("This isn't folder. you can use <cd> to go up folder."); } else if (comd == "paint") { System.CLI.Applications.Paint(); } else if (comd == "changelog") { Console.WriteLine("You are running version 0.0.3 Dev. Only Luka see the DEV!"); Console.WriteLine("v0.0.2:"); Console.WriteLine("Blue screen with cursor. not clearing."); Console.WriteLine("v0.0.3:"); Console.WriteLine("Command Line Shell with Broken CLI, but working unresponsive GUI, but with DOS-like Shell. I/O, Audio, Multithreading, Shutdown doesn't work."); } else if (comd == "help") { Console.WriteLine("Copyright 2015 Thontelix TriangleOS. Special thanks to Cosmos .net ASM Compiler."); Console.WriteLine("type CHANGELOG to get version changes."); Console.WriteLine("How to use:"); Console.WriteLine("After every typed command, press enter. the output of command will be detailed. If there is blinking bottom line cursor, then you need to input something, if it doesn't give any feedback, then its operating a activity. If you want GUI, type 'boot' and press enter."); Console.WriteLine("Commands:"); Console.WriteLine("shutdown - Gives you ability to safe turn off PC"); Console.WriteLine("boot - Boots you into TriangleOS"); Console.WriteLine("reboot - Reboots your PC"); Console.WriteLine("cd - Travel through directories"); Console.WriteLine("dir - Read content of directory"); Console.WriteLine("echo - Backs string you enter"); Console.WriteLine("shutdown - Backs string you enter"); Console.WriteLine("soundtest - Speakers Drivers. Doesnt work for now"); Console.WriteLine("help - Gives you list of commands."); Console.WriteLine("paint - Console Paint App. CAUTION:Not for epilepsy persons."); } else { Console.WriteLine("That command doesn't exist. :("); } } ```
priority
dup tries to pop more stuff from analytical stack than there is log error exception system exception error compiling method systemvoidkernelcommandsinputcommand system exception opcode il dup tries to pop more stuff from analytical stack than there is at cosmos ilopcode interpretstacktypes idictionary aopcodes stack astack boolean asituationchanged amaxrecursiondepth in c users luka desktop cosmos master source cosmos ilopcode cs line at cosmos appassembler interpretinstructionstodeterminestacktypes list acurrentgroup in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler emitinstructions methodinfo amethod list acurrentgroup boolean in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line system exception opcode il dup tries to pop more stuff from analytical stack than there is at cosmos ilopcode interpretstacktypes idictionary aopcodes stack astack boolean asituationchanged amaxrecursiondepth in c users luka desktop cosmos master source cosmos ilopcode cs line at cosmos appassembler interpretinstructionstodeterminestacktypes list acurrentgroup in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler emitinstructions methodinfo amethod list acurrentgroup boolean in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line end of inner exception stack trace at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos ilscanner assemble in c users luka desktop cosmos master source cosmos ilscanner cs line at cosmos ilscanner execute methodbase astartmethod in c users luka desktop cosmos master source cosmos ilscanner cs line at cosmos compilerengine execute in c users luka desktop cosmos master source cosmos compilerengine cs line and there is code where it heappen c public static void inputcommand console write d command comd console readline comd comd tolower if comd reboot h power restart else if comd shutdown h power shutdown else if comd echo console write echo arg console readline console writeline arg else if comd notepad system cli applications notepad else if comd cls console clear console writeline triangleos console writeline else if comd soundtest console write frequency arg console readline console write duration optarg console readline console write eh this isn t implemted right now h multimedia speakers callsound int parse arg int parse optarg else if comd boot console writeline starting triangleos drivers processmanager process audio new processmanager process processmanager process graphics new processmanager process graphics processthread new system threading thread h graphics lowlevel init processmanager process mouse new processmanager process mouse processthread new system threading thread h mouse initmouse graphics start mouse start audio processthread new system threading thread h multimedia speakers intailizeaudio kernel gui else if comd cliboot system cli controls textbox text new system cli controls textbox text x text y text length text drawtextbox text typeinto system cli controls button ok new system cli controls button ok y ok x ok width ok height ok text ok ok drawbutton else if comd calculator system cli applications calculator else if comd cd h graphics console erro impossible operation performed can t request i o while it isn t running else if comd dir console writeline this isn t folder you can use to go up folder else if comd paint system cli applications paint else if comd changelog console writeline you are running version dev only luka see the dev console writeline console writeline blue screen with cursor not clearing console writeline console writeline command line shell with broken cli but working unresponsive gui but with dos like shell i o audio multithreading shutdown doesn t work else if comd help console writeline copyright thontelix triangleos special thanks to cosmos net asm compiler console writeline type changelog to get version changes console writeline how to use console writeline after every typed command press enter the output of command will be detailed if there is blinking bottom line cursor then you need to input something if it doesn t give any feedback then its operating a activity if you want gui type boot and press enter console writeline commands console writeline shutdown gives you ability to safe turn off pc console writeline boot boots you into triangleos console writeline reboot reboots your pc console writeline cd travel through directories console writeline dir read content of directory console writeline echo backs string you enter console writeline shutdown backs string you enter console writeline soundtest speakers drivers doesnt work for now console writeline help gives you list of commands console writeline paint console paint app caution not for epilepsy persons else console writeline that command doesn t exist
1
240,185
7,800,589,404
IssuesEvent
2018-06-09 11:19:53
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0010214: improve calendar performance by reducing the number of recurring events fetched
Calendar Mantis high priority
**Reported by pschuele on 5 Sep 2014 16:16** **Version:** Collin (2013.10.8crowdfunding2) improve calendar performance by reducing the number of recurring events fetched -&gt; because we load all related data for every recurring candidate even if it does not match the period filter
1.0
0010214: improve calendar performance by reducing the number of recurring events fetched - **Reported by pschuele on 5 Sep 2014 16:16** **Version:** Collin (2013.10.8crowdfunding2) improve calendar performance by reducing the number of recurring events fetched -&gt; because we load all related data for every recurring candidate even if it does not match the period filter
priority
improve calendar performance by reducing the number of recurring events fetched reported by pschuele on sep version collin improve calendar performance by reducing the number of recurring events fetched gt because we load all related data for every recurring candidate even if it does not match the period filter
1
82,368
3,605,944,718
IssuesEvent
2016-02-04 08:57:04
TrinityCore/TrinityCore
https://api.github.com/repos/TrinityCore/TrinityCore
closed
[3.3.5] [Script] ICC Emblem of Frost bug
Branch-3.3.5a Comp-Core Priority-High
Dead characters get frost without raid save ! How to reproduce: 1- Need 2 character. create a raid and enter into ICC. 2- Before first boss let's one of characters die and realese spirit. don't resurrect. 3- Kill first boss with another character. 4- Two characters get the frost. but the dead character did'nt get save. and can do it again! This worked for next bosses. rev: c04c409
1.0
[3.3.5] [Script] ICC Emblem of Frost bug - Dead characters get frost without raid save ! How to reproduce: 1- Need 2 character. create a raid and enter into ICC. 2- Before first boss let's one of characters die and realese spirit. don't resurrect. 3- Kill first boss with another character. 4- Two characters get the frost. but the dead character did'nt get save. and can do it again! This worked for next bosses. rev: c04c409
priority
icc emblem of frost bug dead characters get frost without raid save how to reproduce need character create a raid and enter into icc before first boss let s one of characters die and realese spirit don t resurrect kill first boss with another character two characters get the frost but the dead character did nt get save and can do it again this worked for next bosses rev
1
757,617
26,521,577,842
IssuesEvent
2023-01-19 03:30:49
pibolib/hack16-2
https://api.github.com/repos/pibolib/hack16-2
closed
Implement enemy dodge system
enhancement gameplay priority:high
This requires a rewrite to the way player bullets are handled in regards to hurting the enemy. Specifications: 1. Remove direct call to Enemy:take_damage() 2. ~~Create signal on Enemy for bullet_collide~~ **Just calls the collision function directly, no signal needed** 4. ~~On bullet_collide, handle on a case by case (inherited class specific) basis~~ **Handles movement on a case by case basis, dodge() super function to always be called in inherited classes** 6. Always handle dodges before taking damage
1.0
Implement enemy dodge system - This requires a rewrite to the way player bullets are handled in regards to hurting the enemy. Specifications: 1. Remove direct call to Enemy:take_damage() 2. ~~Create signal on Enemy for bullet_collide~~ **Just calls the collision function directly, no signal needed** 4. ~~On bullet_collide, handle on a case by case (inherited class specific) basis~~ **Handles movement on a case by case basis, dodge() super function to always be called in inherited classes** 6. Always handle dodges before taking damage
priority
implement enemy dodge system this requires a rewrite to the way player bullets are handled in regards to hurting the enemy specifications remove direct call to enemy take damage create signal on enemy for bullet collide just calls the collision function directly no signal needed on bullet collide handle on a case by case inherited class specific basis handles movement on a case by case basis dodge super function to always be called in inherited classes always handle dodges before taking damage
1
766,406
26,882,646,626
IssuesEvent
2023-02-05 20:28:24
sczerwinski/wavefront-obj-intellij-plugin
https://api.github.com/repos/sczerwinski/wavefront-obj-intellij-plugin
closed
Objects without material (or with non-existing material) not rendered
type:bug resolution:done priority:high component:3d
### Steps 1. Remove all material references from OBJ file. OR 2. Remove the material used in an OBJ file. ### Result The OBJ file is not rendered. ### Expected Result The OBJ file should be normally rendered, just without a material (default textures/colours).
1.0
Objects without material (or with non-existing material) not rendered - ### Steps 1. Remove all material references from OBJ file. OR 2. Remove the material used in an OBJ file. ### Result The OBJ file is not rendered. ### Expected Result The OBJ file should be normally rendered, just without a material (default textures/colours).
priority
objects without material or with non existing material not rendered steps remove all material references from obj file or remove the material used in an obj file result the obj file is not rendered expected result the obj file should be normally rendered just without a material default textures colours
1
95,990
3,962,917,705
IssuesEvent
2016-05-02 18:33:44
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
closed
Support for Esri Feature Service Symbology
addition: feature priority: high
The Esri Feature Service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend/layer selector. Instead, the viewer should interrogate the service and interpret the Esri symbology values to generate an appropriate graphic. Range, Unique Value and Simple rendering should be supported.
1.0
Support for Esri Feature Service Symbology - The Esri Feature Service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend/layer selector. Instead, the viewer should interrogate the service and interpret the Esri symbology values to generate an appropriate graphic. Range, Unique Value and Simple rendering should be supported.
priority
support for esri feature service symbology the esri feature service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend layer selector instead the viewer should interrogate the service and interpret the esri symbology values to generate an appropriate graphic range unique value and simple rendering should be supported
1
80,367
3,561,152,316
IssuesEvent
2016-01-23 16:13:02
cuckoosandbox/cuckoo
https://api.github.com/repos/cuckoosandbox/cuckoo
closed
Memory dump only readable by root
Bug (to verify) High Priority
We are running cuckoo from git on ubuntu 14.04, and when a memory dump is taken by libvirt the file is only readable by root, which causes problems for volatility. Just creating the file before calling libvirt seems to solve the problem. My ugly hack in cuckoo/lib/common/abstracts.py from line 422: def dump_memory(self, label, path): """Takes a memory dump. @param path: path to where to store the memory dump. """ log.debug("Dumping memory for machine %s", label) #Creating file before call to libvirt to get read permissions for user after dump open(path,'w').write('') conn = self._connect() try: self.vms[label].coreDump(path, flags=libvirt.VIR_DUMP_MEMORY_ONLY) except libvirt.libvirtError as e: raise CuckooMachineError("Error dumping memory virtual machine " "{0}: {1}".format(label, e)) finally: self._disconnect(conn)
1.0
Memory dump only readable by root - We are running cuckoo from git on ubuntu 14.04, and when a memory dump is taken by libvirt the file is only readable by root, which causes problems for volatility. Just creating the file before calling libvirt seems to solve the problem. My ugly hack in cuckoo/lib/common/abstracts.py from line 422: def dump_memory(self, label, path): """Takes a memory dump. @param path: path to where to store the memory dump. """ log.debug("Dumping memory for machine %s", label) #Creating file before call to libvirt to get read permissions for user after dump open(path,'w').write('') conn = self._connect() try: self.vms[label].coreDump(path, flags=libvirt.VIR_DUMP_MEMORY_ONLY) except libvirt.libvirtError as e: raise CuckooMachineError("Error dumping memory virtual machine " "{0}: {1}".format(label, e)) finally: self._disconnect(conn)
priority
memory dump only readable by root we are running cuckoo from git on ubuntu and when a memory dump is taken by libvirt the file is only readable by root which causes problems for volatility just creating the file before calling libvirt seems to solve the problem my ugly hack in cuckoo lib common abstracts py from line def dump memory self label path takes a memory dump param path path to where to store the memory dump log debug dumping memory for machine s label creating file before call to libvirt to get read permissions for user after dump open path w write conn self connect try self vms coredump path flags libvirt vir dump memory only except libvirt libvirterror as e raise cuckoomachineerror error dumping memory virtual machine format label e finally self disconnect conn
1
175,772
6,553,922,830
IssuesEvent
2017-09-06 01:58:47
kinueng/bluetooth-scale
https://api.github.com/repos/kinueng/bluetooth-scale
closed
Start scanning after disconnecting from BLE device
enhancement priority/high status/inprogress
The end goal is to allow the app to always be running, but the BLE device can come and go as it pleases.
1.0
Start scanning after disconnecting from BLE device - The end goal is to allow the app to always be running, but the BLE device can come and go as it pleases.
priority
start scanning after disconnecting from ble device the end goal is to allow the app to always be running but the ble device can come and go as it pleases
1
73,375
3,411,272,070
IssuesEvent
2015-12-05 00:57:49
Ecotrust/COMPASS
https://api.github.com/repos/Ecotrust/COMPASS
closed
info icon on 'Active Tab' layers does nothing
bug High Priority
Either create info-dropdowns (like first tab) or remove icons.
1.0
info icon on 'Active Tab' layers does nothing - Either create info-dropdowns (like first tab) or remove icons.
priority
info icon on active tab layers does nothing either create info dropdowns like first tab or remove icons
1
304,248
9,329,409,585
IssuesEvent
2019-03-28 02:15:41
InQuest/ThreatKB
https://api.github.com/repos/InQuest/ThreatKB
closed
Dynamically update the title.
high-priority
Let the title be "ThreatKB" everywhere, unless, you're viewing or editing an artifact (C2 IP, C2 DNS, YARA). In that case, make the titles: * KB: $IP * KB: $Domain * KB: YARA Signature Name Will help identify the appropriate tabs at a glance, when we have multiple open.
1.0
Dynamically update the title. - Let the title be "ThreatKB" everywhere, unless, you're viewing or editing an artifact (C2 IP, C2 DNS, YARA). In that case, make the titles: * KB: $IP * KB: $Domain * KB: YARA Signature Name Will help identify the appropriate tabs at a glance, when we have multiple open.
priority
dynamically update the title let the title be threatkb everywhere unless you re viewing or editing an artifact ip dns yara in that case make the titles kb ip kb domain kb yara signature name will help identify the appropriate tabs at a glance when we have multiple open
1
793,110
27,983,086,942
IssuesEvent
2023-03-26 11:46:15
AY2223S2-CS2103T-W12-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-W12-1/tp
closed
Update isSamePerson check to check for equality of nric only
type.Bug priority.High severity.Low
Currently, isSamePerson checks for equality of name and nric. This may not match expectations as nric should uniquely identify a person and each person should only have one main name.
1.0
Update isSamePerson check to check for equality of nric only - Currently, isSamePerson checks for equality of name and nric. This may not match expectations as nric should uniquely identify a person and each person should only have one main name.
priority
update issameperson check to check for equality of nric only currently issameperson checks for equality of name and nric this may not match expectations as nric should uniquely identify a person and each person should only have one main name
1
530,370
15,421,828,030
IssuesEvent
2021-03-05 13:36:29
Systems-Learning-and-Development-Lab/MMM
https://api.github.com/repos/Systems-Learning-and-Development-Lab/MMM
closed
Remove existing ball traces if balls are erased
enhancement priority-high
If user clicks on buttons in the UI to remove balls, then traces should be removed as well. ![image](https://user-images.githubusercontent.com/12688282/109966797-56ce6580-7cf9-11eb-8972-8121428acf44.png) In response to [comment](https://github.com/Systems-Learning-and-Development-Lab/MMM/issues/40#issuecomment-789879108)
1.0
Remove existing ball traces if balls are erased - If user clicks on buttons in the UI to remove balls, then traces should be removed as well. ![image](https://user-images.githubusercontent.com/12688282/109966797-56ce6580-7cf9-11eb-8972-8121428acf44.png) In response to [comment](https://github.com/Systems-Learning-and-Development-Lab/MMM/issues/40#issuecomment-789879108)
priority
remove existing ball traces if balls are erased if user clicks on buttons in the ui to remove balls then traces should be removed as well in response to
1
230,200
7,605,351,971
IssuesEvent
2018-04-30 08:35:34
esaude/esaude-emr-poc
https://api.github.com/repos/esaude/esaude-emr-poc
closed
[Access Control] Cannot login with a user with receptionist (admin) role only
High Priority bug
Actual Results -- The system is not permitting the login of a user with receptionist (admin) role only Expected results -- Should allow the login for all the POC roles Steps to reproduce -- In OpenMRS admin page create a user with POC: Receptionist (Admin) role only: Screenshot/Attachment (Optional) -- A visual description of the unexpected behaviour. ![image](https://user-images.githubusercontent.com/3810265/39396085-e95a3cd6-4ae7-11e8-9918-f0fbd13ae2c9.png)
1.0
[Access Control] Cannot login with a user with receptionist (admin) role only - Actual Results -- The system is not permitting the login of a user with receptionist (admin) role only Expected results -- Should allow the login for all the POC roles Steps to reproduce -- In OpenMRS admin page create a user with POC: Receptionist (Admin) role only: Screenshot/Attachment (Optional) -- A visual description of the unexpected behaviour. ![image](https://user-images.githubusercontent.com/3810265/39396085-e95a3cd6-4ae7-11e8-9918-f0fbd13ae2c9.png)
priority
cannot login with a user with receptionist admin role only actual results the system is not permitting the login of a user with receptionist admin role only expected results should allow the login for all the poc roles steps to reproduce in openmrs admin page create a user with poc receptionist admin role only screenshot attachment optional a visual description of the unexpected behaviour
1
82,151
3,603,541,996
IssuesEvent
2016-02-03 19:26:57
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
ListMixin map/extend generic method comments
analyzer-strong-mode area-analyzer priority-high
Right now they don't match the generic method comment signatures on Iterable, so this is an error: ````dart class _FbList<E> extends Object with ListMixin<E> implements List<E> ``` Marked high-pri because this is easy to hit. CC @leafpetersen, as this is related to SDK working with strong mode. (this is sort of area-sdk, but only affects Analyzer strong mode. Not sure best way to label it.)
1.0
ListMixin map/extend generic method comments - Right now they don't match the generic method comment signatures on Iterable, so this is an error: ````dart class _FbList<E> extends Object with ListMixin<E> implements List<E> ``` Marked high-pri because this is easy to hit. CC @leafpetersen, as this is related to SDK working with strong mode. (this is sort of area-sdk, but only affects Analyzer strong mode. Not sure best way to label it.)
priority
listmixin map extend generic method comments right now they don t match the generic method comment signatures on iterable so this is an error dart class fblist extends object with listmixin implements list marked high pri because this is easy to hit cc leafpetersen as this is related to sdk working with strong mode this is sort of area sdk but only affects analyzer strong mode not sure best way to label it
1
690,549
23,663,955,685
IssuesEvent
2022-08-26 18:35:04
tellor-io/telliot-feeds
https://api.github.com/repos/tellor-io/telliot-feeds
opened
Fix & add DIVA Protocol integration tests
enhancement priority:high
- finish tests in `tests/integrations/diva_protocol/test_report.py` - finish tests in `tests/integrations/diva_protocol/test_e2e_create_report_settle_pool.py`
1.0
Fix & add DIVA Protocol integration tests - - finish tests in `tests/integrations/diva_protocol/test_report.py` - finish tests in `tests/integrations/diva_protocol/test_e2e_create_report_settle_pool.py`
priority
fix add diva protocol integration tests finish tests in tests integrations diva protocol test report py finish tests in tests integrations diva protocol test create report settle pool py
1
810,845
30,264,331,181
IssuesEvent
2023-07-07 10:35:18
zeek/zeek
https://api.github.com/repos/zeek/zeek
closed
Incomplete Zeek-side type generation for exported Spicy unit with unit switch
Priority: High Area: Spicy
The following test should work, but fails since the Zeek-side field for the conditionally filled field `foo` is not generated. ```ruby # @TEST-EXEC: spicyz -do test.hlto test.spicy test.evt # @TEST-EXEC: zeek -Cr ${TRACES}/udp-packet.pcap test.hlto main.zeek >output 2>&1 # @TEST-EXEC: btest-diff output # @TEST-START-FILE test.spicy module TEST; public type Message = unit { switch (1) { * -> foo: bytes &eod; }; }; # @TEST-END-FILE # @TEST-START-FILE test.evt import TEST; protocol analyzer spicy::Test over UDP: port 0/udp - 42000/udp, parse with TEST::Message; export TEST::Message; on TEST::Message -> event TEST_ZEEK::MessageEvt(self); # @TEST-END-FILE # @TEST-START-FILE main.zeek event TEST_ZEEK::MessageEvt(message: TEST::Message) { } # @TEST-END-FILE ``` As of `6.1.0-dev.152` (ccfc68fccd9b0635210bf951ee6d4769f997d157) this emits the following runtime warning: ``` XXXXXXXXXX.XXXXXX analyzer error in <...>/test.evt, line 8: Event parameter mismatch, no matching record field for field 'foo' ``` This looks similar to #3083 with the additional twist of a unit switch.
1.0
Incomplete Zeek-side type generation for exported Spicy unit with unit switch - The following test should work, but fails since the Zeek-side field for the conditionally filled field `foo` is not generated. ```ruby # @TEST-EXEC: spicyz -do test.hlto test.spicy test.evt # @TEST-EXEC: zeek -Cr ${TRACES}/udp-packet.pcap test.hlto main.zeek >output 2>&1 # @TEST-EXEC: btest-diff output # @TEST-START-FILE test.spicy module TEST; public type Message = unit { switch (1) { * -> foo: bytes &eod; }; }; # @TEST-END-FILE # @TEST-START-FILE test.evt import TEST; protocol analyzer spicy::Test over UDP: port 0/udp - 42000/udp, parse with TEST::Message; export TEST::Message; on TEST::Message -> event TEST_ZEEK::MessageEvt(self); # @TEST-END-FILE # @TEST-START-FILE main.zeek event TEST_ZEEK::MessageEvt(message: TEST::Message) { } # @TEST-END-FILE ``` As of `6.1.0-dev.152` (ccfc68fccd9b0635210bf951ee6d4769f997d157) this emits the following runtime warning: ``` XXXXXXXXXX.XXXXXX analyzer error in <...>/test.evt, line 8: Event parameter mismatch, no matching record field for field 'foo' ``` This looks similar to #3083 with the additional twist of a unit switch.
priority
incomplete zeek side type generation for exported spicy unit with unit switch the following test should work but fails since the zeek side field for the conditionally filled field foo is not generated ruby test exec spicyz do test hlto test spicy test evt test exec zeek cr traces udp packet pcap test hlto main zeek output test exec btest diff output test start file test spicy module test public type message unit switch foo bytes eod test end file test start file test evt import test protocol analyzer spicy test over udp port udp udp parse with test message export test message on test message event test zeek messageevt self test end file test start file main zeek event test zeek messageevt message test message test end file as of dev this emits the following runtime warning xxxxxxxxxx xxxxxx analyzer error in test evt line event parameter mismatch no matching record field for field foo this looks similar to with the additional twist of a unit switch
1
190,145
6,810,323,568
IssuesEvent
2017-11-05 03:59:03
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
closed
[Fork-Join] Cannot add the a fork join
0.95 Priority/Highest Severity/Blocker
Pack: https://wso2.org/jenkins/view/All%20Builds/job/ballerinalang/job/tools-distribution/880/org.ballerinalang.tools$ballerina-tools/ Cannot add a fork join as shown in the attached image ![com_0 94_10](https://user-images.githubusercontent.com/15624590/32402832-206d0bba-c152-11e7-9572-fa4346b3f6c7.png)
1.0
[Fork-Join] Cannot add the a fork join - Pack: https://wso2.org/jenkins/view/All%20Builds/job/ballerinalang/job/tools-distribution/880/org.ballerinalang.tools$ballerina-tools/ Cannot add a fork join as shown in the attached image ![com_0 94_10](https://user-images.githubusercontent.com/15624590/32402832-206d0bba-c152-11e7-9572-fa4346b3f6c7.png)
priority
cannot add the a fork join pack cannot add a fork join as shown in the attached image
1
134,267
5,222,927,919
IssuesEvent
2017-01-27 10:18:58
hpi-swt2/workshop-portal
https://api.github.com/repos/hpi-swt2/workshop-portal
closed
Custom variable application fields in application
High Priority needs acceptance team-helene
- [ ] render fields on application fill-out page **Dependency:** #322
1.0
Custom variable application fields in application - - [ ] render fields on application fill-out page **Dependency:** #322
priority
custom variable application fields in application render fields on application fill out page dependency
1
162,886
6,177,832,534
IssuesEvent
2017-07-02 05:53:22
redox-os/ion
https://api.github.com/repos/redox-os/ion
closed
Incomplete Signal Handling
enhancement high-priority
- [x] Implement SIGINT Handling - [x] Implement SIGTERM Handling - [ ] Ensure that SIGPIPE is working correctly - [X] Implement SIGSTOP Handling - [X] Implement SIGCONT Handling - [x] Implement `bg` - [ ] Implement `fg` - [ ] Implement SIGHUP Handling - [ ] disown - [ ] suspend
1.0
Incomplete Signal Handling - - [x] Implement SIGINT Handling - [x] Implement SIGTERM Handling - [ ] Ensure that SIGPIPE is working correctly - [X] Implement SIGSTOP Handling - [X] Implement SIGCONT Handling - [x] Implement `bg` - [ ] Implement `fg` - [ ] Implement SIGHUP Handling - [ ] disown - [ ] suspend
priority
incomplete signal handling implement sigint handling implement sigterm handling ensure that sigpipe is working correctly implement sigstop handling implement sigcont handling implement bg implement fg implement sighup handling disown suspend
1
651,642
21,484,116,610
IssuesEvent
2022-04-26 20:56:49
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
[Bug] Crash : Ankidroid crashes on checking media.
Priority-High Bug Reproduced
###### Reproduction Steps - Updated from 2.15.6 (Play store ver.) to 2.16alpha58 and when i opened a card i started getting "failed to load : failed to load ...." ![Screenshot_2022-04-16-17-34-50-744_com.ichi2.anki.png](https://user-images.githubusercontent.com/63903377/163674249-884f95df-96c1-44f1-8a0e-8a5fdc0edefd.png) - So i decided to use check media which instantly crashed the app. - After that i decided to force sync from anki web to ankidroid which also crashed the app. - Decided to check if my library was corrupted for some reason or not by downloading a parallel of 2.15.6 on which it seemed to work perfectly, updated the parallel to alpha58 and it was also crashing. - The alpha also doesn't seem to be detecting the backups neither is it creating them properly from what i can tell. ![Screenshot_2022-04-16-17-29-38-032_ru.zdevs.zarchiver.pro.png](https://user-images.githubusercontent.com/63903377/163674472-24c1d8ce-3c83-4239-b7cb-ed7f6b001482.png) The only one which was good on 16/04 was made by the stable parallel. - Just to confirm if the problem was from this alpha or from before this I decided to test on alpha 56 parallel and got the same issues uninstalled the parallel and installed alpha 55 parallel and there wasn't a crash this time, decided to update 55 to 56 and test again and there wasn't a crash, same with from 56 to 58. - ~~Btw I did try to reinstall main alpha58 directly when i was originally getting the crash before the above point happened and the crashes still happened.~~ I thought i did but seems like i hadn't installed it directly and when I just did it seems to work fine on installing directly and this only happens if you update directly from 2.15.6 to 2.16 Alpha56/58/59. - Have confirmed that it doesn't crash anymore on main after uninstalling 58 and installing 55 then 56 and 58 again after that. - ~~Backup still don't seem to be being created properly but am not sure yet, so will confirm after I am.~~ It is detecting backups again and the backups are being created properly. - Also half of this happened while writing this so it might seem like a story. ###### Expected Result App checks media , directly syncs from anki web to ankidroid , Backup being created properly. ###### Actual Result App crashes if you have updated directly to alpha 58 or installing alpha58 directly over current library. ###### Debug info AnkiDroid Version = 2.16alpha58 Android Version = 11 Manufacturer = Xiaomi Model = Redmi Note 9 Pro Max Hardware = qcom Webview User Agent = Mozilla/5.0 (Linux; Android 11; Redmi Note 9 Pro Max Build/RKQ1.200826.002; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/100.0.4896.88 Mobile Safari/537.36 ACRA UUID = d9199133-a62c-4572-ad4f-899c4d0b5307 Scheduler = std2 Crash Reports Enabled = true DatabaseV2 Enabled = true ###### Research *Enter an [x] character to confirm the points below:* - [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid - [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue - [x] I have searched for similar existing issues here and on the user forum - [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
1.0
[Bug] Crash : Ankidroid crashes on checking media. - ###### Reproduction Steps - Updated from 2.15.6 (Play store ver.) to 2.16alpha58 and when i opened a card i started getting "failed to load : failed to load ...." ![Screenshot_2022-04-16-17-34-50-744_com.ichi2.anki.png](https://user-images.githubusercontent.com/63903377/163674249-884f95df-96c1-44f1-8a0e-8a5fdc0edefd.png) - So i decided to use check media which instantly crashed the app. - After that i decided to force sync from anki web to ankidroid which also crashed the app. - Decided to check if my library was corrupted for some reason or not by downloading a parallel of 2.15.6 on which it seemed to work perfectly, updated the parallel to alpha58 and it was also crashing. - The alpha also doesn't seem to be detecting the backups neither is it creating them properly from what i can tell. ![Screenshot_2022-04-16-17-29-38-032_ru.zdevs.zarchiver.pro.png](https://user-images.githubusercontent.com/63903377/163674472-24c1d8ce-3c83-4239-b7cb-ed7f6b001482.png) The only one which was good on 16/04 was made by the stable parallel. - Just to confirm if the problem was from this alpha or from before this I decided to test on alpha 56 parallel and got the same issues uninstalled the parallel and installed alpha 55 parallel and there wasn't a crash this time, decided to update 55 to 56 and test again and there wasn't a crash, same with from 56 to 58. - ~~Btw I did try to reinstall main alpha58 directly when i was originally getting the crash before the above point happened and the crashes still happened.~~ I thought i did but seems like i hadn't installed it directly and when I just did it seems to work fine on installing directly and this only happens if you update directly from 2.15.6 to 2.16 Alpha56/58/59. - Have confirmed that it doesn't crash anymore on main after uninstalling 58 and installing 55 then 56 and 58 again after that. - ~~Backup still don't seem to be being created properly but am not sure yet, so will confirm after I am.~~ It is detecting backups again and the backups are being created properly. - Also half of this happened while writing this so it might seem like a story. ###### Expected Result App checks media , directly syncs from anki web to ankidroid , Backup being created properly. ###### Actual Result App crashes if you have updated directly to alpha 58 or installing alpha58 directly over current library. ###### Debug info AnkiDroid Version = 2.16alpha58 Android Version = 11 Manufacturer = Xiaomi Model = Redmi Note 9 Pro Max Hardware = qcom Webview User Agent = Mozilla/5.0 (Linux; Android 11; Redmi Note 9 Pro Max Build/RKQ1.200826.002; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/100.0.4896.88 Mobile Safari/537.36 ACRA UUID = d9199133-a62c-4572-ad4f-899c4d0b5307 Scheduler = std2 Crash Reports Enabled = true DatabaseV2 Enabled = true ###### Research *Enter an [x] character to confirm the points below:* - [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid - [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue - [x] I have searched for similar existing issues here and on the user forum - [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
priority
crash ankidroid crashes on checking media reproduction steps updated from play store ver to and when i opened a card i started getting failed to load failed to load so i decided to use check media which instantly crashed the app after that i decided to force sync from anki web to ankidroid which also crashed the app decided to check if my library was corrupted for some reason or not by downloading a parallel of on which it seemed to work perfectly updated the parallel to and it was also crashing the alpha also doesn t seem to be detecting the backups neither is it creating them properly from what i can tell the only one which was good on was made by the stable parallel just to confirm if the problem was from this alpha or from before this i decided to test on alpha parallel and got the same issues uninstalled the parallel and installed alpha parallel and there wasn t a crash this time decided to update to and test again and there wasn t a crash same with from to btw i did try to reinstall main directly when i was originally getting the crash before the above point happened and the crashes still happened i thought i did but seems like i hadn t installed it directly and when i just did it seems to work fine on installing directly and this only happens if you update directly from to have confirmed that it doesn t crash anymore on main after uninstalling and installing then and again after that backup still don t seem to be being created properly but am not sure yet so will confirm after i am it is detecting backups again and the backups are being created properly also half of this happened while writing this so it might seem like a story expected result app checks media directly syncs from anki web to ankidroid backup being created properly actual result app crashes if you have updated directly to alpha or installing directly over current library debug info ankidroid version android version manufacturer xiaomi model redmi note pro max hardware qcom webview user agent mozilla linux android redmi note pro max build wv applewebkit khtml like gecko version chrome mobile safari acra uuid scheduler crash reports enabled true enabled true research enter an character to confirm the points below i have read the and am reporting a bug or enhancement request specific to ankidroid i have checked the and the and could not find a solution to my issue i have searched for similar existing issues here and on the user forum optional i have confirmed the issue is not resolved in the latest alpha release
1
468,035
13,460,220,533
IssuesEvent
2020-09-09 13:20:52
UniversityOfHelsinkiCS/lomake
https://api.github.com/repos/UniversityOfHelsinkiCS/lomake
closed
Create downloadable csv export for system wide answers
enhancement high priority
Format should be something like this: | Programme name| Answer 1 | Answer 2| |----------|:-------------:|------:| | prog 1 | answer 1 | answer 2 | | prog 2 | answer 1 | answer 2 | | prog 3 | answer 1 | answer 2 |
1.0
Create downloadable csv export for system wide answers - Format should be something like this: | Programme name| Answer 1 | Answer 2| |----------|:-------------:|------:| | prog 1 | answer 1 | answer 2 | | prog 2 | answer 1 | answer 2 | | prog 3 | answer 1 | answer 2 |
priority
create downloadable csv export for system wide answers format should be something like this programme name answer answer prog answer answer prog answer answer prog answer answer
1
380,656
11,269,287,171
IssuesEvent
2020-01-14 08:34:36
wazuh/wazuh-kibana-app
https://api.github.com/repos/wazuh/wazuh-kibana-app
closed
Wazuh FIM - "show files" timeout when monitor lost of files
bug help wanted priority/high
| Wazuh | Elastic | Rev | | ----- | ------- | --- | | 3.9.2 | 7.1.1 | 444 | **Description** I cannot use show files when lots of files is monitored with FIM. I increased the timeout in the kibana configs, but i got timeout after 20 sec, or 'Maximum call stack size exceeded'. My question: - how i increase the API timeout? **Steps to reproduce** 1. Go to 'Integrity monitoring' 2. Click on 'show files' 3. see error 'Maximum call stack size exceeded' 4. search with "wp" 5. see the file list 6. click on "Last modified ^ " 7. see error "Wazuh API error: 3021 - Timeout executing API request" **Screenshots** ![timeout](https://user-images.githubusercontent.com/44770411/60963500-4215b080-a311-11e9-86ff-d4d0d44ef7e2.png) ![maximum](https://user-images.githubusercontent.com/44770411/60963503-43df7400-a311-11e9-899e-eb8d6f26a7fe.png) **Additional context** I use docker to run manager kibana.yml : - elasticsearch.requestTimeout: 900000 kibana wazuh plugin: - timout: 90000 the request time after 20sec I monitor several WP directory. I don't monitor the cache files.
1.0
Wazuh FIM - "show files" timeout when monitor lost of files - | Wazuh | Elastic | Rev | | ----- | ------- | --- | | 3.9.2 | 7.1.1 | 444 | **Description** I cannot use show files when lots of files is monitored with FIM. I increased the timeout in the kibana configs, but i got timeout after 20 sec, or 'Maximum call stack size exceeded'. My question: - how i increase the API timeout? **Steps to reproduce** 1. Go to 'Integrity monitoring' 2. Click on 'show files' 3. see error 'Maximum call stack size exceeded' 4. search with "wp" 5. see the file list 6. click on "Last modified ^ " 7. see error "Wazuh API error: 3021 - Timeout executing API request" **Screenshots** ![timeout](https://user-images.githubusercontent.com/44770411/60963500-4215b080-a311-11e9-86ff-d4d0d44ef7e2.png) ![maximum](https://user-images.githubusercontent.com/44770411/60963503-43df7400-a311-11e9-899e-eb8d6f26a7fe.png) **Additional context** I use docker to run manager kibana.yml : - elasticsearch.requestTimeout: 900000 kibana wazuh plugin: - timout: 90000 the request time after 20sec I monitor several WP directory. I don't monitor the cache files.
priority
wazuh fim show files timeout when monitor lost of files wazuh elastic rev description i cannot use show files when lots of files is monitored with fim i increased the timeout in the kibana configs but i got timeout after sec or maximum call stack size exceeded my question how i increase the api timeout steps to reproduce go to integrity monitoring click on show files see error maximum call stack size exceeded search with wp see the file list click on last modified see error wazuh api error timeout executing api request screenshots additional context i use docker to run manager kibana yml elasticsearch requesttimeout kibana wazuh plugin timout the request time after i monitor several wp directory i don t monitor the cache files
1
658,418
21,892,053,182
IssuesEvent
2022-05-20 03:30:36
TeamSparker/Spark-Server
https://api.github.com/repos/TeamSparker/Spark-Server
opened
[FIX] recordDB.getFeedRecordsByRoomIds UTC + 9 처리
🦋 영권 High-Priority 🚨 🚨 ISSUE 🚨
## 📌 Issue - CURRENT_DATE - INTERVAL '7 days' 가 UTC + 9 기준이 아님 ## 📝 To-do - [ ] Fix it
1.0
[FIX] recordDB.getFeedRecordsByRoomIds UTC + 9 처리 - ## 📌 Issue - CURRENT_DATE - INTERVAL '7 days' 가 UTC + 9 기준이 아님 ## 📝 To-do - [ ] Fix it
priority
recorddb getfeedrecordsbyroomids utc 처리 📌 issue current date interval days 가 utc 기준이 아님 📝 to do fix it
1
804,757
29,499,641,081
IssuesEvent
2023-06-02 20:17:31
culibraries/library-reporting
https://api.github.com/repos/culibraries/library-reporting
closed
Trace Lost Reports
High Priority
Reports that are ran using a combination of status and item notes to check for missing items before they are declared missing. Items to Export: Title List Filter - [ ] Note field = "xxx" Sierra Query: ITEM STATUS equal to "l" AND (ITEM LOCATION equal to "norst" OR ITEM LOCATION equal to "norov" OR ITEM LOCATION equal to "asnst" OR ITEM LOCATION equal to "asnov" OR ITEM LOCATION equal to "artst" OR ITEM LOCATION equal to "artov" OR ITEM LOCATION equal to "scist" OR ITEM LOCATION equal to "sciov") AND ITEM NOTE All Fields don't have "march 2022"
1.0
Trace Lost Reports - Reports that are ran using a combination of status and item notes to check for missing items before they are declared missing. Items to Export: Title List Filter - [ ] Note field = "xxx" Sierra Query: ITEM STATUS equal to "l" AND (ITEM LOCATION equal to "norst" OR ITEM LOCATION equal to "norov" OR ITEM LOCATION equal to "asnst" OR ITEM LOCATION equal to "asnov" OR ITEM LOCATION equal to "artst" OR ITEM LOCATION equal to "artov" OR ITEM LOCATION equal to "scist" OR ITEM LOCATION equal to "sciov") AND ITEM NOTE All Fields don't have "march 2022"
priority
trace lost reports reports that are ran using a combination of status and item notes to check for missing items before they are declared missing items to export title list filter note field xxx sierra query item status equal to l and item location equal to norst or item location equal to norov or item location equal to asnst or item location equal to asnov or item location equal to artst or item location equal to artov or item location equal to scist or item location equal to sciov and item note all fields don t have march
1
577,199
17,105,124,842
IssuesEvent
2021-07-09 16:29:43
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Segmentation fault when a Tensor backward hook removes itself
high priority module: autograd module: crash triaged
## 🐛 Bug If a Tensor backward hook removes itself during execution *and* returns a modified gradient, PyTorch will segfault. ## To Reproduce ```python import torch def install_hook(tensor): handle = None def hook(tensor): handle.remove() return torch.zeros_like(tensor) handle = tensor.register_hook(hook) def test_hook(): t = torch.ones((1, 5)) t.requires_grad_() install_hook(t) (t ** 2).mean().backward() print(t.grad) if __name__ == '__main__': test_hook() ``` ``` $ python test_hook.py Segmentation fault (core dumped) ``` Stacktrace, as recorded by gdb: ``` #0 func_get_name (op=0x7fffb186b790, _unused_ignored=0x0) at /tmp/build/80754af9/python_1599203911753/work/Objects/funcobject.c:299 #1 0x0000555555692f65 in _PyObject_GenericGetAttrWithDict (obj=0x7fffb186b790, name=0x7ffff74b1770, dict=0x0, suppress=0) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:1254 #2 0x00005555556a90fb in PyObject_GetAttrString (v=0x7fffb186b790, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:846 #3 0x00005555555e5beb in PyObject_HasAttrString (v=<optimized out>, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:854 #4 0x00007fffdc92377b in hook_name(_object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #5 0x00007fffdc923dcc in torch::autograd::PyFunctionPreHook::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #6 0x00007fffd60cb1ff in torch::autograd::AccumulateGrad::apply(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #7 0x00007fffd60c651b in torch::autograd::Node::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #8 0x00007fffd60c1dc3 in torch::autograd::Engine::evaluate_function(std::shared_ptr<torch::autograd::GraphTask>&, torch::autograd::Node*, torch::autograd::InputBuffer&, std::shared_ptr<torch::autograd::ReadyQueue> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #9 0x00007fffd60c2698 in torch::autograd::Engine::thread_main(std::shared_ptr<torch::autograd::GraphTask> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #10 0x00007fffd60bd481 in torch::autograd::Engine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #11 0x00007fffdc9177c1 in torch::autograd::python::PythonEngine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #12 0x00007fffd60bf375 in torch::autograd::Engine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #13 0x00007fffdc917711 in torch::autograd::python::PythonEngine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #14 0x00007fffdc91818c in THPEngine_run_backward(_object*, _object*, _object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #15 0x00005555556d3f76 in cfunction_call_varargs (kwargs=<optimized out>, args=<optimized out>, func=0x7ffff74b58b0) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:742 #16 PyCFunction_Call (func=0x7ffff74b58b0, args=<optimized out>, kwargs=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:772 #17 0x000055555569185f in _PyObject_MakeTpCall (callable=0x7ffff74b58b0, args=<optimized out>, nargs=<optimized out>, keywords=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:159 #18 0x0000555555715b90 in _PyObject_Vectorcall (kwnames=0x7fffb6e94400, nargsf=<optimized out>, args=<optimized out>, callable=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:125 #19 call_function (kwnames=0x7fffb6e94400, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #20 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515 #21 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb2384660, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x7fffb6e2de98, kwargs=0x7ffff75807c8, kwcount=<optimized out>, kwstep=1, defs=0x7fffb23883c8, defcount=5, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7ffff74ae7f0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #22 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75807a8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435 #23 0x000055555565411a in _PyObject_Vectorcall (kwnames=0x7fffb6e2de80, nargsf=<optimized out>, args=<optimized out>, callable=0x7fffb2355ca0) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #24 call_function (kwnames=0x7fffb6e2de80, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #25 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515 #26 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb6e35450, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x7ffff752e3c0, kwcount=<optimized out>, kwstep=1, defs=0x7fffb6e39aa8, defcount=4, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7fffb6f83970) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #27 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff752e3b8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435 #28 0x000055555565475e in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff752e3b8, callable=0x7fffb6d64790) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #29 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #30 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3486 #31 0x00005555556df86b in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:283 #32 _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75bc5b0, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:410 #33 0x0000555555652b84 in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff75bc5b0, callable=0x7fffb18c3c10) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #34 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #35 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3500 #36 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7ffff74943a0, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x0, kwcount=<optimized out>, kwstep=2, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #37 0x00005555556df754 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4327 #38 0x000055555576dedc in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:718 #39 0x000055555576df84 in run_eval_code_obj (co=0x7ffff74943a0, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1125 #40 0x00005555557a01f4 in run_mod (mod=<optimized out>, filename=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, flags=<optimized out>, arena=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1147 #41 0x00005555556686e1 in PyRun_FileExFlags (fp=0x5555558eaf90, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1063 #42 0x0000555555668ac6 in PyRun_SimpleFileExFlags (fp=0x5555558eaf90, filename=<optimized out>, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:428 #43 0x000055555566998b in pymain_run_file (cf=0x7fffffffdbd8, config=0x5555558eb5b0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:387 #44 pymain_run_python (exitcode=0x7fffffffdbd0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:612 #45 Py_RunMain () at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:691 #46 0x00005555557a2d19 in Py_BytesMain (argc=<optimized out>, argv=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:1137 #47 0x00007ffff7db00b3 in __libc_start_main (main=0x55555566a460 <main>, argc=2, argv=0x7fffffffddd8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffddc8) at ../csu/libc-start.c:308 #48 0x0000555555732e93 in _start () at ../sysdeps/x86_64/elf/start.S:103 ``` ## Expected behavior It doesn't segfault. If a hook removing itself is considered an error, it should raise an exception. ## Environment ``` PyTorch version: 1.9.0a0+git6af4e99 Is debug build: False CUDA used to build PyTorch: 10.1 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.2 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: 13.0.0-++20210418052640+d480f968ad8b-1~exp1~20210418153358.383 CMake version: version 3.18.2 Python version: 3.8 (64-bit runtime) Is CUDA available: True CUDA runtime version: 10.1.243 GPU models and configuration: GPU 0: GeForce RTX 2080 SUPER Nvidia driver version: 450.119.03 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.2 [pip3] numpydoc==1.1.0 [pip3] torch==1.9.0a0+git9def776 [conda] blas 1.0 mkl [conda] mkl 2020.2 256 [conda] mkl-include 2021.2.0 h06a4308_296 [conda] mkl-service 2.3.0 py38he904b0f_0 [conda] mkl_fft 1.2.0 py38h23d657b_0 [conda] mkl_random 1.1.1 py38h0573a6f_0 [conda] numpy 1.19.2 py38h54aff64_0 [conda] numpy-base 1.19.2 py38hfa32c7d_0 [conda] numpydoc 1.1.0 pyhd3eb1b0_1 [conda] torch 1.9.0a0+git9def776 pypi_0 pypi ``` ## Additional context Tested on PyTorch built from master as of 9def776cd6d775779db9839d942a01b4b082175c as well as on 1.8.1 cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @gqchen @pearu @nikitaved @soulitzer @Lezcano
1.0
Segmentation fault when a Tensor backward hook removes itself - ## 🐛 Bug If a Tensor backward hook removes itself during execution *and* returns a modified gradient, PyTorch will segfault. ## To Reproduce ```python import torch def install_hook(tensor): handle = None def hook(tensor): handle.remove() return torch.zeros_like(tensor) handle = tensor.register_hook(hook) def test_hook(): t = torch.ones((1, 5)) t.requires_grad_() install_hook(t) (t ** 2).mean().backward() print(t.grad) if __name__ == '__main__': test_hook() ``` ``` $ python test_hook.py Segmentation fault (core dumped) ``` Stacktrace, as recorded by gdb: ``` #0 func_get_name (op=0x7fffb186b790, _unused_ignored=0x0) at /tmp/build/80754af9/python_1599203911753/work/Objects/funcobject.c:299 #1 0x0000555555692f65 in _PyObject_GenericGetAttrWithDict (obj=0x7fffb186b790, name=0x7ffff74b1770, dict=0x0, suppress=0) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:1254 #2 0x00005555556a90fb in PyObject_GetAttrString (v=0x7fffb186b790, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:846 #3 0x00005555555e5beb in PyObject_HasAttrString (v=<optimized out>, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:854 #4 0x00007fffdc92377b in hook_name(_object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #5 0x00007fffdc923dcc in torch::autograd::PyFunctionPreHook::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #6 0x00007fffd60cb1ff in torch::autograd::AccumulateGrad::apply(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #7 0x00007fffd60c651b in torch::autograd::Node::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #8 0x00007fffd60c1dc3 in torch::autograd::Engine::evaluate_function(std::shared_ptr<torch::autograd::GraphTask>&, torch::autograd::Node*, torch::autograd::InputBuffer&, std::shared_ptr<torch::autograd::ReadyQueue> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #9 0x00007fffd60c2698 in torch::autograd::Engine::thread_main(std::shared_ptr<torch::autograd::GraphTask> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #10 0x00007fffd60bd481 in torch::autograd::Engine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #11 0x00007fffdc9177c1 in torch::autograd::python::PythonEngine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #12 0x00007fffd60bf375 in torch::autograd::Engine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so #13 0x00007fffdc917711 in torch::autograd::python::PythonEngine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #14 0x00007fffdc91818c in THPEngine_run_backward(_object*, _object*, _object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so #15 0x00005555556d3f76 in cfunction_call_varargs (kwargs=<optimized out>, args=<optimized out>, func=0x7ffff74b58b0) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:742 #16 PyCFunction_Call (func=0x7ffff74b58b0, args=<optimized out>, kwargs=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:772 #17 0x000055555569185f in _PyObject_MakeTpCall (callable=0x7ffff74b58b0, args=<optimized out>, nargs=<optimized out>, keywords=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:159 #18 0x0000555555715b90 in _PyObject_Vectorcall (kwnames=0x7fffb6e94400, nargsf=<optimized out>, args=<optimized out>, callable=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:125 #19 call_function (kwnames=0x7fffb6e94400, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #20 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515 #21 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb2384660, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x7fffb6e2de98, kwargs=0x7ffff75807c8, kwcount=<optimized out>, kwstep=1, defs=0x7fffb23883c8, defcount=5, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7ffff74ae7f0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #22 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75807a8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435 #23 0x000055555565411a in _PyObject_Vectorcall (kwnames=0x7fffb6e2de80, nargsf=<optimized out>, args=<optimized out>, callable=0x7fffb2355ca0) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #24 call_function (kwnames=0x7fffb6e2de80, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #25 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515 #26 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb6e35450, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x7ffff752e3c0, kwcount=<optimized out>, kwstep=1, defs=0x7fffb6e39aa8, defcount=4, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7fffb6f83970) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #27 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff752e3b8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435 #28 0x000055555565475e in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff752e3b8, callable=0x7fffb6d64790) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #29 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #30 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3486 #31 0x00005555556df86b in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:283 #32 _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75bc5b0, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:410 #33 0x0000555555652b84 in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff75bc5b0, callable=0x7fffb18c3c10) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127 #34 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963 #35 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3500 #36 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7ffff74943a0, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x0, kwcount=<optimized out>, kwstep=2, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298 #37 0x00005555556df754 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4327 #38 0x000055555576dedc in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:718 #39 0x000055555576df84 in run_eval_code_obj (co=0x7ffff74943a0, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1125 #40 0x00005555557a01f4 in run_mod (mod=<optimized out>, filename=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, flags=<optimized out>, arena=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1147 #41 0x00005555556686e1 in PyRun_FileExFlags (fp=0x5555558eaf90, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1063 #42 0x0000555555668ac6 in PyRun_SimpleFileExFlags (fp=0x5555558eaf90, filename=<optimized out>, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:428 #43 0x000055555566998b in pymain_run_file (cf=0x7fffffffdbd8, config=0x5555558eb5b0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:387 #44 pymain_run_python (exitcode=0x7fffffffdbd0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:612 #45 Py_RunMain () at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:691 #46 0x00005555557a2d19 in Py_BytesMain (argc=<optimized out>, argv=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:1137 #47 0x00007ffff7db00b3 in __libc_start_main (main=0x55555566a460 <main>, argc=2, argv=0x7fffffffddd8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffddc8) at ../csu/libc-start.c:308 #48 0x0000555555732e93 in _start () at ../sysdeps/x86_64/elf/start.S:103 ``` ## Expected behavior It doesn't segfault. If a hook removing itself is considered an error, it should raise an exception. ## Environment ``` PyTorch version: 1.9.0a0+git6af4e99 Is debug build: False CUDA used to build PyTorch: 10.1 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.2 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: 13.0.0-++20210418052640+d480f968ad8b-1~exp1~20210418153358.383 CMake version: version 3.18.2 Python version: 3.8 (64-bit runtime) Is CUDA available: True CUDA runtime version: 10.1.243 GPU models and configuration: GPU 0: GeForce RTX 2080 SUPER Nvidia driver version: 450.119.03 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.2 [pip3] numpydoc==1.1.0 [pip3] torch==1.9.0a0+git9def776 [conda] blas 1.0 mkl [conda] mkl 2020.2 256 [conda] mkl-include 2021.2.0 h06a4308_296 [conda] mkl-service 2.3.0 py38he904b0f_0 [conda] mkl_fft 1.2.0 py38h23d657b_0 [conda] mkl_random 1.1.1 py38h0573a6f_0 [conda] numpy 1.19.2 py38h54aff64_0 [conda] numpy-base 1.19.2 py38hfa32c7d_0 [conda] numpydoc 1.1.0 pyhd3eb1b0_1 [conda] torch 1.9.0a0+git9def776 pypi_0 pypi ``` ## Additional context Tested on PyTorch built from master as of 9def776cd6d775779db9839d942a01b4b082175c as well as on 1.8.1 cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @gqchen @pearu @nikitaved @soulitzer @Lezcano
priority
segmentation fault when a tensor backward hook removes itself 🐛 bug if a tensor backward hook removes itself during execution and returns a modified gradient pytorch will segfault to reproduce python import torch def install hook tensor handle none def hook tensor handle remove return torch zeros like tensor handle tensor register hook hook def test hook t torch ones t requires grad install hook t t mean backward print t grad if name main test hook python test hook py segmentation fault core dumped stacktrace as recorded by gdb func get name op unused ignored at tmp build python work objects funcobject c in pyobject genericgetattrwithdict obj name dict suppress at tmp build python work objects object c in pyobject getattrstring v name at tmp build python work objects object c in pyobject hasattrstring v name at tmp build python work objects object c in hook name object from home nelhage sw lib site packages torch lib libtorch python so in torch autograd pyfunctionprehook operator std vector const from home nelhage sw lib site packages torch lib libtorch python so in torch autograd accumulategrad apply std vector from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd node operator std vector from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine evaluate function std shared ptr torch autograd node torch autograd inputbuffer std shared ptr const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine thread main std shared ptr const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine execute with graph task std shared ptr const std shared ptr torch autograd inputbuffer from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd python pythonengine execute with graph task std shared ptr const std shared ptr torch autograd inputbuffer from home nelhage sw lib site packages torch lib libtorch python so in torch autograd engine execute std vector const std vector const bool bool bool std vector const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd python pythonengine execute std vector const std vector const bool bool bool std vector const from home nelhage sw lib site packages torch lib libtorch python so in thpengine run backward object object object from home nelhage sw lib site packages torch lib libtorch python so in cfunction call varargs kwargs args func at tmp build python work objects call c pycfunction call func args kwargs at tmp build python work objects call c in pyobject maketpcall callable args nargs keywords at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in function code fastcall globals nargs args co at tmp build python work objects call c pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at tmp build python work python ceval c in pyeval evalcode co globals locals at tmp build python work python ceval c in run eval code obj co globals locals at tmp build python work python pythonrun c in run mod mod filename globals locals flags arena at tmp build python work python pythonrun c in pyrun fileexflags fp filename str start globals locals closeit flags at tmp build python work python pythonrun c in pyrun simplefileexflags fp filename closeit flags at tmp build python work python pythonrun c in pymain run file cf config at tmp build python work modules main c pymain run python exitcode at tmp build python work modules main c py runmain at tmp build python work modules main c in py bytesmain argc argv at tmp build python work modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start at sysdeps elf start s expected behavior it doesn t segfault if a hook removing itself is considered an error it should raise an exception environment pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version cmake version version python version bit runtime is cuda available true cuda runtime version gpu models and configuration gpu geforce rtx super nvidia driver version cudnn version could not collect hip runtime version n a miopen runtime version n a versions of relevant libraries numpy numpydoc torch blas mkl mkl mkl include mkl service mkl fft mkl random numpy numpy base numpydoc torch pypi pypi additional context tested on pytorch built from master as of as well as on cc ezyang gchanan bdhirsh jbschlosser alband gqchen pearu nikitaved soulitzer lezcano
1
321,998
9,811,221,873
IssuesEvent
2019-06-12 22:49:27
keystonejs/keystone-5
https://api.github.com/repos/keystonejs/keystone-5
closed
isRequired has no effect
bug graphql high priority schema
``` js fields: { description: { type: Text, isRequired: true } } ``` Expectation is that it fails if `null` (or `''`, aka empty string) in the administration UI on creation. Actuality: It creates an item without an error. **edit:** Unclear if this is meant be enforced by GraphQL, but, I can verify that you can `create*` `isRequired` fields as `null` without error too. `isUnique` **is** enforced by GraphQL.
1.0
isRequired has no effect - ``` js fields: { description: { type: Text, isRequired: true } } ``` Expectation is that it fails if `null` (or `''`, aka empty string) in the administration UI on creation. Actuality: It creates an item without an error. **edit:** Unclear if this is meant be enforced by GraphQL, but, I can verify that you can `create*` `isRequired` fields as `null` without error too. `isUnique` **is** enforced by GraphQL.
priority
isrequired has no effect js fields description type text isrequired true expectation is that it fails if null or aka empty string in the administration ui on creation actuality it creates an item without an error edit unclear if this is meant be enforced by graphql but i can verify that you can create isrequired fields as null without error too isunique is enforced by graphql
1
674,578
23,058,357,111
IssuesEvent
2022-07-25 07:38:17
zitadel/zitadel
https://api.github.com/repos/zitadel/zitadel
closed
[OAuth]: add more claims to introspection response
category: docs type: enhancement lang: go category: backend priority: high
RFC7662 Token Introspection https://datatracker.ietf.org/doc/html/rfc7662#section-2.2 specifies that only claim `active` is required to be returned. Additionally we also return `scope`, `client_id`, `username` and all requested claims according to userinfo response (incl. `sub`, `iss`) We should also add `token_type`, `exp`, `iat`, `nbf`, `aud`, `jti` This would also allow libraries to cache the response more efficiently. Requires changes to the OIDC Library. **Acceptance criteria** - [x] the claims above are returned on the response - [x] documentation
1.0
[OAuth]: add more claims to introspection response - RFC7662 Token Introspection https://datatracker.ietf.org/doc/html/rfc7662#section-2.2 specifies that only claim `active` is required to be returned. Additionally we also return `scope`, `client_id`, `username` and all requested claims according to userinfo response (incl. `sub`, `iss`) We should also add `token_type`, `exp`, `iat`, `nbf`, `aud`, `jti` This would also allow libraries to cache the response more efficiently. Requires changes to the OIDC Library. **Acceptance criteria** - [x] the claims above are returned on the response - [x] documentation
priority
add more claims to introspection response token introspection specifies that only claim active is required to be returned additionally we also return scope client id username and all requested claims according to userinfo response incl sub iss we should also add token type exp iat nbf aud jti this would also allow libraries to cache the response more efficiently requires changes to the oidc library acceptance criteria the claims above are returned on the response documentation
1
346,289
10,410,355,708
IssuesEvent
2019-09-13 11:09:52
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
conan create -> conan upload
complex: medium priority: high stage: queue type: look into
It will create a file with the reference created (or the install reference if specified). It will be a pref (without revision). Together with #5196 would alleviate nicely the `conan create` -> `conan upload` typical CI flow that now requires painful parsings. @solvingj
1.0
conan create -> conan upload - It will create a file with the reference created (or the install reference if specified). It will be a pref (without revision). Together with #5196 would alleviate nicely the `conan create` -> `conan upload` typical CI flow that now requires painful parsings. @solvingj
priority
conan create conan upload it will create a file with the reference created or the install reference if specified it will be a pref without revision together with would alleviate nicely the conan create conan upload typical ci flow that now requires painful parsings solvingj
1
448,563
12,952,690,279
IssuesEvent
2020-07-19 21:31:22
passwall/passwall-desktop
https://api.github.com/repos/passwall/passwall-desktop
closed
Parse CSV data to array
Priority: High Status: Available Type: Enhancement
**Is your feature request related to a problem? Please describe.** Now import function is added ass seen in the picture. User should import csv files with this. <img width="533" alt="Screen Shot 2020-07-15 at 23 03 36" src="https://user-images.githubusercontent.com/3696314/87590503-8d665a00-c6ef-11ea-9bb4-ade42c6682c2.png"> I am going to encrypt and add values to database however I need someone to parse this csv content line by line as array. The import handle function is in **/src/renderer/App.vue** like this: ```js onImport() { dialog.showOpenDialog({ properties: ['openFile'] }, files => { if (files.length === 0) return try { const fileContent = fs.readFileSync(files[0]).toString() // TODO: Parse fileContent (CSV data) to array console.log(fileContent) } catch (error) { console.log(error) } }) } ``` **Describe the solution you'd like** Example CSV content is here https://github.com/passwall/passwall-server/blob/master/example.csv Can someone parse and give me the values of CSV lines as array and javascript variables? Thank you.
1.0
Parse CSV data to array - **Is your feature request related to a problem? Please describe.** Now import function is added ass seen in the picture. User should import csv files with this. <img width="533" alt="Screen Shot 2020-07-15 at 23 03 36" src="https://user-images.githubusercontent.com/3696314/87590503-8d665a00-c6ef-11ea-9bb4-ade42c6682c2.png"> I am going to encrypt and add values to database however I need someone to parse this csv content line by line as array. The import handle function is in **/src/renderer/App.vue** like this: ```js onImport() { dialog.showOpenDialog({ properties: ['openFile'] }, files => { if (files.length === 0) return try { const fileContent = fs.readFileSync(files[0]).toString() // TODO: Parse fileContent (CSV data) to array console.log(fileContent) } catch (error) { console.log(error) } }) } ``` **Describe the solution you'd like** Example CSV content is here https://github.com/passwall/passwall-server/blob/master/example.csv Can someone parse and give me the values of CSV lines as array and javascript variables? Thank you.
priority
parse csv data to array is your feature request related to a problem please describe now import function is added ass seen in the picture user should import csv files with this img width alt screen shot at src i am going to encrypt and add values to database however i need someone to parse this csv content line by line as array the import handle function is in src renderer app vue like this js onimport dialog showopendialog properties files if files length return try const filecontent fs readfilesync files tostring todo parse filecontent csv data to array console log filecontent catch error console log error describe the solution you d like example csv content is here can someone parse and give me the values of csv lines as array and javascript variables thank you
1
390,006
11,520,593,648
IssuesEvent
2020-02-14 15:05:31
openforis/collect-earth-online
https://api.github.com/repos/openforis/collect-earth-online
opened
Allow users to edit their institution's imagery sources (and make sure they appear on the Review Project page)
Feature Request High Priority
There is no way to edit the imagery source. The process of adding WMS services is very error prone (e.g. mistyping the JSON parameters or changing the resolution or format) so you should be able to edit the imagery sources (maybe the base ones like Bing Maps or Digital Globe should be " locked" ) http://i.imgur.com/0jW6gaw.png When I choose the PNOA as a basemap it works fine assessing plots, but if I go back to the project page (using the Edit button) the basemap parameter appears empty, even though PNOA is in the list of avaliable sources. see images. http://i.imgur.com/zurwJEF.png http://i.imgur.com/gPUGd6u.png
1.0
Allow users to edit their institution's imagery sources (and make sure they appear on the Review Project page) - There is no way to edit the imagery source. The process of adding WMS services is very error prone (e.g. mistyping the JSON parameters or changing the resolution or format) so you should be able to edit the imagery sources (maybe the base ones like Bing Maps or Digital Globe should be " locked" ) http://i.imgur.com/0jW6gaw.png When I choose the PNOA as a basemap it works fine assessing plots, but if I go back to the project page (using the Edit button) the basemap parameter appears empty, even though PNOA is in the list of avaliable sources. see images. http://i.imgur.com/zurwJEF.png http://i.imgur.com/gPUGd6u.png
priority
allow users to edit their institution s imagery sources and make sure they appear on the review project page there is no way to edit the imagery source the process of adding wms services is very error prone e g mistyping the json parameters or changing the resolution or format so you should be able to edit the imagery sources maybe the base ones like bing maps or digital globe should be locked when i choose the pnoa as a basemap it works fine assessing plots but if i go back to the project page using the edit button the basemap parameter appears empty even though pnoa is in the list of avaliable sources see images
1
463,579
13,284,356,671
IssuesEvent
2020-08-24 06:06:28
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1731] Crash with Crane
Category: Gameplay Priority: High Status: Fixed
Step to reproduce: - place a new Crane. - take non-constucted blocks from stockpile (I took Birch log): ![image](https://user-images.githubusercontent.com/45708377/90604890-3419ab00-e206-11ea-914c-9197ef1a6a27.png) - start to build: ![image](https://user-images.githubusercontent.com/45708377/90604918-3ed44000-e206-11ea-8c92-60246b9444fe.png) - press LMB - Crash: ``` System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Gameplay.Components.CraneToolComponent.Build(Player player, Vector3i min, Vector3i max)</size> ``` [Crane Crash.txt](https://github.com/StrangeLoopGames/EcoIssues/files/5094658/Crane.Crash.txt) If you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine.
1.0
[0.9.0 staging-1731] Crash with Crane - Step to reproduce: - place a new Crane. - take non-constucted blocks from stockpile (I took Birch log): ![image](https://user-images.githubusercontent.com/45708377/90604890-3419ab00-e206-11ea-914c-9197ef1a6a27.png) - start to build: ![image](https://user-images.githubusercontent.com/45708377/90604918-3ed44000-e206-11ea-8c92-60246b9444fe.png) - press LMB - Crash: ``` System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Gameplay.Components.CraneToolComponent.Build(Player player, Vector3i min, Vector3i max)</size> ``` [Crane Crash.txt](https://github.com/StrangeLoopGames/EcoIssues/files/5094658/Crane.Crash.txt) If you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine.
priority
crash with crane step to reproduce place a new crane take non constucted blocks from stockpile i took birch log start to build press lmb crash system nullreferenceexception object reference not set to an instance of an object at eco gameplay components cranetoolcomponent build player player min max if you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine
1
382,811
11,320,176,831
IssuesEvent
2020-01-21 02:56:40
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
App crash on creating a non-program requisition
Docs: not needed Effort: small Module: indicators Priority: high Priority: release blocker
## Describe the bug App crashes when creating a non-program requisition. ### To reproduce 1. Create a non-program requisition 4. See error ### Expected behaviour No crash ### Proposed Solution Think it's the state init method ### Version and device info - App version: 3.1.1 development - Tablet model: - OS version: ### Additional context <img width="1357" alt="image" src="https://user-images.githubusercontent.com/35858975/72573200-b87f1500-3929-11ea-80df-6d194305a805.png">
2.0
App crash on creating a non-program requisition - ## Describe the bug App crashes when creating a non-program requisition. ### To reproduce 1. Create a non-program requisition 4. See error ### Expected behaviour No crash ### Proposed Solution Think it's the state init method ### Version and device info - App version: 3.1.1 development - Tablet model: - OS version: ### Additional context <img width="1357" alt="image" src="https://user-images.githubusercontent.com/35858975/72573200-b87f1500-3929-11ea-80df-6d194305a805.png">
priority
app crash on creating a non program requisition describe the bug app crashes when creating a non program requisition to reproduce create a non program requisition see error expected behaviour no crash proposed solution think it s the state init method version and device info app version development tablet model os version additional context img width alt image src
1
162,350
6,151,454,059
IssuesEvent
2017-06-28 02:38:25
mercadopago/px-ios
https://api.github.com/repos/mercadopago/px-ios
closed
Salimos de flujo al volver para atrás de revisa y confirma
Priority: High
Deberiamos volver a la ultima pantalla del flujo
1.0
Salimos de flujo al volver para atrás de revisa y confirma - Deberiamos volver a la ultima pantalla del flujo
priority
salimos de flujo al volver para atrás de revisa y confirma deberiamos volver a la ultima pantalla del flujo
1
351,273
10,514,775,904
IssuesEvent
2019-09-28 03:32:43
webhintio/hint
https://api.github.com/repos/webhintio/hint
closed
[Bug] Need to add instructions when clicking on the extension to let users know they should go to the dev tools
area:browser-extension priority:high type:bug
Clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab. This is how it looks like right now: ![image](https://user-images.githubusercontent.com/606594/65796791-60c26100-e122-11e9-8519-5c860f727957.png)
1.0
[Bug] Need to add instructions when clicking on the extension to let users know they should go to the dev tools - Clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab. This is how it looks like right now: ![image](https://user-images.githubusercontent.com/606594/65796791-60c26100-e122-11e9-8519-5c860f727957.png)
priority
need to add instructions when clicking on the extension to let users know they should go to the dev tools clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab this is how it looks like right now
1
768,966
26,989,126,236
IssuesEvent
2023-02-09 18:22:14
repobee/repobee
https://api.github.com/repos/repobee/repobee
closed
Add periodic CI job that ensures the installer can run on the current release
priority:high ci/cd
As noted in #1116, v3.8.0 could not be installed from source (and thus not with the installer). Why the install script is installing from source instead of the wheel on PyPi is a bit of a mystery in itself. Anyway, there is obviously a need for a CI job that ensures that the installer can run according to the install instructions, so this doesn't happen again.
1.0
Add periodic CI job that ensures the installer can run on the current release - As noted in #1116, v3.8.0 could not be installed from source (and thus not with the installer). Why the install script is installing from source instead of the wheel on PyPi is a bit of a mystery in itself. Anyway, there is obviously a need for a CI job that ensures that the installer can run according to the install instructions, so this doesn't happen again.
priority
add periodic ci job that ensures the installer can run on the current release as noted in could not be installed from source and thus not with the installer why the install script is installing from source instead of the wheel on pypi is a bit of a mystery in itself anyway there is obviously a need for a ci job that ensures that the installer can run according to the install instructions so this doesn t happen again
1
670,794
22,704,259,426
IssuesEvent
2022-07-05 13:26:04
skyportal/skyportal
https://api.github.com/repos/skyportal/skyportal
closed
Automating email and SMS notification given specific requirements
high priority
Right now, we can set allowEmailAlerts and allowSMSAlert (in addition to slack notifications). Best I can tell, allowEmailAlerts and allowSMSAlerts do not do anything, and their behavior should be merged with the slack infrastructure (where different types of notifications are allowed).
1.0
Automating email and SMS notification given specific requirements - Right now, we can set allowEmailAlerts and allowSMSAlert (in addition to slack notifications). Best I can tell, allowEmailAlerts and allowSMSAlerts do not do anything, and their behavior should be merged with the slack infrastructure (where different types of notifications are allowed).
priority
automating email and sms notification given specific requirements right now we can set allowemailalerts and allowsmsalert in addition to slack notifications best i can tell allowemailalerts and allowsmsalerts do not do anything and their behavior should be merged with the slack infrastructure where different types of notifications are allowed
1
463,763
13,300,066,025
IssuesEvent
2020-08-25 10:44:20
GluuFederation/oxAuth
https://api.github.com/repos/GluuFederation/oxAuth
closed
If software_statement is present in dynamic registration request, plain claims are lost
bug high priority
## Describe the issue If software_statement is present in dynamic registration request, plain claims are lost. From RFC 7591: ```Client metadata values MAY also be provided in a software statement, as described in Section 2.3 ... If the server supports software statements, client metadata values conveyed in the software statement MUST take precedence over those conveyed using plain JSON elements.``` ## Expected behavior Plain text claims should be present and take part in registration, however those claims that are present in software statement should take higher priority (override plain claims). ## Actual behavior Only software statement claims take part in registration. Plain claims are ignored.
1.0
If software_statement is present in dynamic registration request, plain claims are lost - ## Describe the issue If software_statement is present in dynamic registration request, plain claims are lost. From RFC 7591: ```Client metadata values MAY also be provided in a software statement, as described in Section 2.3 ... If the server supports software statements, client metadata values conveyed in the software statement MUST take precedence over those conveyed using plain JSON elements.``` ## Expected behavior Plain text claims should be present and take part in registration, however those claims that are present in software statement should take higher priority (override plain claims). ## Actual behavior Only software statement claims take part in registration. Plain claims are ignored.
priority
if software statement is present in dynamic registration request plain claims are lost describe the issue if software statement is present in dynamic registration request plain claims are lost from rfc client metadata values may also be provided in a software statement as described in section if the server supports software statements client metadata values conveyed in the software statement must take precedence over those conveyed using plain json elements expected behavior plain text claims should be present and take part in registration however those claims that are present in software statement should take higher priority override plain claims actual behavior only software statement claims take part in registration plain claims are ignored
1
785,958
27,629,658,333
IssuesEvent
2023-03-10 09:53:11
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Make release compatibility filtering opt-in
type:feature priority-2-high breaking status:in-progress v35
### What would you like Renovate to be able to do? Default to no release filtering (today only python exact versions is supported) so that we can add more advanced filtering in future. ### If you have any ideas on how this should be implemented, please tell us here. I'd like to make it an enum in case we want to have different degrees of filtering in future. Feature name `releaseConstraintFiltering`, default value `"none"`. Alternative value `"strict"`, and put this code behind an `if (config.releaseConstraintFiltering === 'strict')`: https://github.com/renovatebot/renovate/blob/8e8cbba0e62ac5ccac0a24de295a468c5c2101b8/lib/modules/datasource/index.ts#L398-L419 ### Is this a feature you are interested in implementing yourself? Maybe
1.0
Make release compatibility filtering opt-in - ### What would you like Renovate to be able to do? Default to no release filtering (today only python exact versions is supported) so that we can add more advanced filtering in future. ### If you have any ideas on how this should be implemented, please tell us here. I'd like to make it an enum in case we want to have different degrees of filtering in future. Feature name `releaseConstraintFiltering`, default value `"none"`. Alternative value `"strict"`, and put this code behind an `if (config.releaseConstraintFiltering === 'strict')`: https://github.com/renovatebot/renovate/blob/8e8cbba0e62ac5ccac0a24de295a468c5c2101b8/lib/modules/datasource/index.ts#L398-L419 ### Is this a feature you are interested in implementing yourself? Maybe
priority
make release compatibility filtering opt in what would you like renovate to be able to do default to no release filtering today only python exact versions is supported so that we can add more advanced filtering in future if you have any ideas on how this should be implemented please tell us here i d like to make it an enum in case we want to have different degrees of filtering in future feature name releaseconstraintfiltering default value none alternative value strict and put this code behind an if config releaseconstraintfiltering strict is this a feature you are interested in implementing yourself maybe
1
792,441
27,960,524,707
IssuesEvent
2023-03-24 15:20:26
KDT3-Final-6/final-project-FE
https://api.github.com/repos/KDT3-Final-6/final-project-FE
opened
Design: 로그인 & 회원가입 마크업
Status: Available Status: Review Needed Priority: High
## ✔️ 체크사항 - [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요. ## 💡 개발 목적 - 로그인 & 회원가입 마크업 ## 🌐 세부 내용 - [ ] 회원가입 - [ ] 로그인 ## ⚙️ 참조 [피그마] (https://www.figma.com/file/AVUhg8ZkRtoas58FMo95SE/%EA%B3%A0%ED%88%AC%EA%B2%8C%EB%8D%94-6%EC%A1%B0?node-id=1-2&t=PdJGv1AtnUhZyXyp-0)
1.0
Design: 로그인 & 회원가입 마크업 - ## ✔️ 체크사항 - [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요. ## 💡 개발 목적 - 로그인 & 회원가입 마크업 ## 🌐 세부 내용 - [ ] 회원가입 - [ ] 로그인 ## ⚙️ 참조 [피그마] (https://www.figma.com/file/AVUhg8ZkRtoas58FMo95SE/%EA%B3%A0%ED%88%AC%EA%B2%8C%EB%8D%94-6%EC%A1%B0?node-id=1-2&t=PdJGv1AtnUhZyXyp-0)
priority
design 로그인 회원가입 마크업 ✔️ 체크사항 제목은 개발 페이지 개발 목적 으로 작성해 주세요 💡 개발 목적 로그인 회원가입 마크업 🌐 세부 내용 회원가입 로그인 ⚙️ 참조
1
520,057
15,077,801,727
IssuesEvent
2021-02-05 07:38:57
wso2/cellery-integration-tests
https://api.github.com/repos/wso2/cellery-integration-tests
closed
Package structure should be revisited and refactored
Priority/High Resolution/Won’t Fix Type/Improvement
**Description:** The current code has no proper structure and it makes it hard maintain the code. The packaging should be restructured.
1.0
Package structure should be revisited and refactored - **Description:** The current code has no proper structure and it makes it hard maintain the code. The packaging should be restructured.
priority
package structure should be revisited and refactored description the current code has no proper structure and it makes it hard maintain the code the packaging should be restructured
1
522,761
15,167,424,299
IssuesEvent
2021-02-12 17:46:18
ooni/ooni.org
https://api.github.com/repos/ooni/ooni.org
closed
Update OONI Data Policy
effort/M priority/high website
I'll be updating the OONI Data Policy to reflect the changes made in the latest OONI Probe mobile and desktop app releases. Specifically, I'll be listing all the analytics tools we're using across platforms, and editing other sections now that we have removed the option to opt-in to IP address collection (and other changes).
1.0
Update OONI Data Policy - I'll be updating the OONI Data Policy to reflect the changes made in the latest OONI Probe mobile and desktop app releases. Specifically, I'll be listing all the analytics tools we're using across platforms, and editing other sections now that we have removed the option to opt-in to IP address collection (and other changes).
priority
update ooni data policy i ll be updating the ooni data policy to reflect the changes made in the latest ooni probe mobile and desktop app releases specifically i ll be listing all the analytics tools we re using across platforms and editing other sections now that we have removed the option to opt in to ip address collection and other changes
1
740,828
25,769,876,131
IssuesEvent
2022-12-09 06:52:06
makeniii/NBA-Elo-Rankings
https://api.github.com/repos/makeniii/NBA-Elo-Rankings
closed
2. Get PR of all NBA teams
Priority: High
Returns a list of all NBA teams sorted by PR. Acceptance criteria: - [x] Returns a list of all NBA teams - [x] For each NBA team - - [x] Have name of team - - [x] Current PR - [x] Sort list by descending PR
1.0
2. Get PR of all NBA teams - Returns a list of all NBA teams sorted by PR. Acceptance criteria: - [x] Returns a list of all NBA teams - [x] For each NBA team - - [x] Have name of team - - [x] Current PR - [x] Sort list by descending PR
priority
get pr of all nba teams returns a list of all nba teams sorted by pr acceptance criteria returns a list of all nba teams for each nba team have name of team current pr sort list by descending pr
1
490,859
14,141,031,197
IssuesEvent
2020-11-10 12:05:50
ansible/awx
https://api.github.com/repos/ansible/awx
closed
[ui_next] Surveys cannot be enabled via Survey Toggle for WFJT
component:ui_next priority:high qe:escape qe:regression state:in_progress type:bug
##### ISSUE TYPE - Bug Report ##### SUMMARY Surveys cannot be enabled via Survey Toggle for WFJT ##### ENVIRONMENT * AWX version: 13 ##### STEPS TO REPRODUCE 1. Create a WFJT 2. Navigate to Survey tab and create a survey 3. Click 'Enable Surveys toggle' Note: Error message ##### EXPECTED RESULTS Survey toggled on ##### ACTUAL RESULTS Survey not toggled on Error message ##### ADDITIONAL INFORMATION
1.0
[ui_next] Surveys cannot be enabled via Survey Toggle for WFJT - ##### ISSUE TYPE - Bug Report ##### SUMMARY Surveys cannot be enabled via Survey Toggle for WFJT ##### ENVIRONMENT * AWX version: 13 ##### STEPS TO REPRODUCE 1. Create a WFJT 2. Navigate to Survey tab and create a survey 3. Click 'Enable Surveys toggle' Note: Error message ##### EXPECTED RESULTS Survey toggled on ##### ACTUAL RESULTS Survey not toggled on Error message ##### ADDITIONAL INFORMATION
priority
surveys cannot be enabled via survey toggle for wfjt issue type bug report summary surveys cannot be enabled via survey toggle for wfjt environment awx version steps to reproduce create a wfjt navigate to survey tab and create a survey click enable surveys toggle note error message expected results survey toggled on actual results survey not toggled on error message additional information
1
557,265
16,505,076,636
IssuesEvent
2021-05-25 18:14:54
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
opened
regularly run tests on current release
priority/high type/bug
We should be able to know the same day we start encountering issues like https://github.com/airbytehq/airbyte/issues/3588 that stop fresh installs. We shouldn't need to wait until a user reports it.
1.0
regularly run tests on current release - We should be able to know the same day we start encountering issues like https://github.com/airbytehq/airbyte/issues/3588 that stop fresh installs. We shouldn't need to wait until a user reports it.
priority
regularly run tests on current release we should be able to know the same day we start encountering issues like that stop fresh installs we shouldn t need to wait until a user reports it
1
178,755
6,618,291,944
IssuesEvent
2017-09-21 07:30:56
hobbit-project/platform
https://api.github.com/repos/hobbit-project/platform
opened
Logout not working for anonymous user
component: UI priority: high type: bug
## Problem When logged in as anonymous user, the logout button is not working correctly. Instead of being redirected to the login page of Keycloak the user is redirected to the main page of the GUI still logged in as anonymous user.
1.0
Logout not working for anonymous user - ## Problem When logged in as anonymous user, the logout button is not working correctly. Instead of being redirected to the login page of Keycloak the user is redirected to the main page of the GUI still logged in as anonymous user.
priority
logout not working for anonymous user problem when logged in as anonymous user the logout button is not working correctly instead of being redirected to the login page of keycloak the user is redirected to the main page of the gui still logged in as anonymous user
1
714,401
24,560,393,126
IssuesEvent
2022-10-12 19:41:29
interaction-lab/MoveToCode
https://api.github.com/repos/interaction-lab/MoveToCode
closed
Behavior Clean Up
high priority
- [x] high five needs sound - [x] sound for behaviors - [x] switch mode flashing when should be off - [x] remove piece back to text - [x] don't do anything on connect/update action at end of loggable action - [x] pulse next maze only on high five - [x] enter goal explode only on correct - [x] put high five down if press next maze button - [x] high five no work on end every time - [x] hold picture above head - [x] kuri head busted - [x] ask for help button??? - [x] if kuri colliding eith maze, move away --- - [x] move then point should be in sequence - [x] bug on start up run away? -> might just be the ipad... - [x] look at user on high five - [x] text too large in front of me - [x] turn off kuri move when not editor - [x] too close to stuff - [x] on reset, reset all anims - [x] choose side of checker hit - [x] say something code help - [x] move away 1 metere from user - [x] code help should move to an obj - [x] help button stopped working - [x] add connectors hit as a human action - [x] build then code text change on maze correct - [x] logger runs a ton, not sure why, need to look into this as it is what is stalling out --- - [x] test out with people --- ### Probably not going to do until playtesting done - [x] add in state where nothing is happening and deal with high five being put down - [x] auto clear long text low prio in `KuriTextManager.cs`
1.0
Behavior Clean Up - - [x] high five needs sound - [x] sound for behaviors - [x] switch mode flashing when should be off - [x] remove piece back to text - [x] don't do anything on connect/update action at end of loggable action - [x] pulse next maze only on high five - [x] enter goal explode only on correct - [x] put high five down if press next maze button - [x] high five no work on end every time - [x] hold picture above head - [x] kuri head busted - [x] ask for help button??? - [x] if kuri colliding eith maze, move away --- - [x] move then point should be in sequence - [x] bug on start up run away? -> might just be the ipad... - [x] look at user on high five - [x] text too large in front of me - [x] turn off kuri move when not editor - [x] too close to stuff - [x] on reset, reset all anims - [x] choose side of checker hit - [x] say something code help - [x] move away 1 metere from user - [x] code help should move to an obj - [x] help button stopped working - [x] add connectors hit as a human action - [x] build then code text change on maze correct - [x] logger runs a ton, not sure why, need to look into this as it is what is stalling out --- - [x] test out with people --- ### Probably not going to do until playtesting done - [x] add in state where nothing is happening and deal with high five being put down - [x] auto clear long text low prio in `KuriTextManager.cs`
priority
behavior clean up high five needs sound sound for behaviors switch mode flashing when should be off remove piece back to text don t do anything on connect update action at end of loggable action pulse next maze only on high five enter goal explode only on correct put high five down if press next maze button high five no work on end every time hold picture above head kuri head busted ask for help button if kuri colliding eith maze move away move then point should be in sequence bug on start up run away might just be the ipad look at user on high five text too large in front of me turn off kuri move when not editor too close to stuff on reset reset all anims choose side of checker hit say something code help move away metere from user code help should move to an obj help button stopped working add connectors hit as a human action build then code text change on maze correct logger runs a ton not sure why need to look into this as it is what is stalling out test out with people probably not going to do until playtesting done add in state where nothing is happening and deal with high five being put down auto clear long text low prio in kuritextmanager cs
1
782,447
27,496,913,941
IssuesEvent
2023-03-05 08:38:05
ThomUK/SPCreporter
https://api.github.com/repos/ThomUK/SPCreporter
opened
Implement a new unit type: 'rate per 1000'
enhancement high priority
Some proportion measures need reporting as a rate per 1000, not as a percentage. Implement rounding for the header row, for both the target, and the last data point
1.0
Implement a new unit type: 'rate per 1000' - Some proportion measures need reporting as a rate per 1000, not as a percentage. Implement rounding for the header row, for both the target, and the last data point
priority
implement a new unit type rate per some proportion measures need reporting as a rate per not as a percentage implement rounding for the header row for both the target and the last data point
1
133,984
5,218,019,273
IssuesEvent
2017-01-26 15:29:41
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
closed
Turn edd_get_payment into a caching layer for EDD_Payment, and allow false on no-payment.
Bug Payments Priority: High
When passing a `$payment_id` that isn't a valid payment, `EDD_Payment::setup_payment()` returns false before setting up the payment. `edd_update_payment_status()` assumes that instantiating `EDD_Payment` returns a valid `EDD_Payment` class, and continues with that assumption, calling `EDD_Payment::save()` no matter the return value. Related to https://trello.com/c/j7LgHA80/436-gravity-forms-checkout-for-easy-digital-downloads#comment-575edeabeb0ad751c227c4e9
1.0
Turn edd_get_payment into a caching layer for EDD_Payment, and allow false on no-payment. - When passing a `$payment_id` that isn't a valid payment, `EDD_Payment::setup_payment()` returns false before setting up the payment. `edd_update_payment_status()` assumes that instantiating `EDD_Payment` returns a valid `EDD_Payment` class, and continues with that assumption, calling `EDD_Payment::save()` no matter the return value. Related to https://trello.com/c/j7LgHA80/436-gravity-forms-checkout-for-easy-digital-downloads#comment-575edeabeb0ad751c227c4e9
priority
turn edd get payment into a caching layer for edd payment and allow false on no payment when passing a payment id that isn t a valid payment edd payment setup payment returns false before setting up the payment edd update payment status assumes that instantiating edd payment returns a valid edd payment class and continues with that assumption calling edd payment save no matter the return value related to
1
364,377
10,763,338,767
IssuesEvent
2019-11-01 03:32:15
ShabadOS/desktop
https://api.github.com/repos/ShabadOS/desktop
closed
feat(frontend/settings): add highlight current line option
Priority: High Status: Confirmed Type: Feature
With the introduction of multiple previous+next lines, it can be confusing to determine what the current line activated is. I suggest adding a "{highlight, accent, contrasting} current line" option with some CSS it would be easier to know what the main line is. I can provide the CSS on a theme-by-theme basis if the react component is set up properly.
1.0
feat(frontend/settings): add highlight current line option - With the introduction of multiple previous+next lines, it can be confusing to determine what the current line activated is. I suggest adding a "{highlight, accent, contrasting} current line" option with some CSS it would be easier to know what the main line is. I can provide the CSS on a theme-by-theme basis if the react component is set up properly.
priority
feat frontend settings add highlight current line option with the introduction of multiple previous next lines it can be confusing to determine what the current line activated is i suggest adding a highlight accent contrasting current line option with some css it would be easier to know what the main line is i can provide the css on a theme by theme basis if the react component is set up properly
1
673,772
23,030,396,796
IssuesEvent
2022-07-22 13:24:59
canonical/prometheus-k8s-operator
https://api.github.com/repos/canonical/prometheus-k8s-operator
closed
MetricsEndpointProvider units are not always reachable via fqdn
Priority: High Type: Bug Area: prometheus_scrape
### Bug Description #293 replaced `network.bind_address()` with `socket.getfqdn()`. This works for in-model relations but breaks for cross-cluster relations. cc: @mateoflorido @stonepreston ### To Reproduce Form a cross-cluster prometheus_scrape relation b/w a machine charm and the prometheus-k8s charm. ### Environment > we have been trying to integrate the prometheus-k8s charm with kube-ovn charm. But it seems that the library gathers the instance-id of our charm instead of the IP address of the unit, therefore metrics are not being collected in the process. > > they are in different models. We have a CK cluster that has the kube-ovn charm as it's CNI charm. > > Kube-OVN is a subordinate machine charm. > -- @mateoflorido ### Relevant log output ```shell ~$ curl 10.152.183.72:9090/api/v1/targets { "status": "success", "data": { "activeTargets": [ { "discoveredLabels": { "__address__": "juju-865c1a-5:10665", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/4", }, "labels": { "instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/4", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/4", }, "scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "scrapeUrl": "http://juju-865c1a-5:10665/metrics", "globalUrl": "http://juju-865c1a-5:10665/metrics", "lastError": 'Get "http://juju-865c1a-5:10665/metrics": dial tcp: lookup juju-865c1a-5 on 10.152.183.213:53: no such host', "lastScrape": "2022-07-15T15:27:59.615508445Z", "lastScrapeDuration": 0.005049553, "health": "down", }, { "discoveredLabels": { "__address__": "juju-865c1a-8:10665", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/0", }, "labels": { "instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/0", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/0", }, "scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "scrapeUrl": "http://juju-865c1a-8:10665/metrics", "globalUrl": "http://juju-865c1a-8:10665/metrics", "lastError": 'Get "http://juju-865c1a-8:10665/metrics": dial tcp: lookup juju-865c1a-8 on 10.152.183.213:53: no such host', "lastScrape": "2022-07-15T15:28:30.262574207Z", "lastScrapeDuration": 0.00596523, "health": "down", }, { "discoveredLabels": { "__address__": "localhost:9090", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "prometheus", }, "labels": {"instance": "localhost:9090", "job": "prometheus"}, "scrapePool": "prometheus", "scrapeUrl": "http://localhost:9090/metrics", "globalUrl": "http://192.168.0.17:9090/metrics", "lastError": "", "lastScrape": "2022-07-15T15:28:46.369824673Z", "lastScrapeDuration": 0.006802522, "health": "up", }, ], "droppedTargets": [], }, } ``` ### Additional context _No response_
1.0
MetricsEndpointProvider units are not always reachable via fqdn - ### Bug Description #293 replaced `network.bind_address()` with `socket.getfqdn()`. This works for in-model relations but breaks for cross-cluster relations. cc: @mateoflorido @stonepreston ### To Reproduce Form a cross-cluster prometheus_scrape relation b/w a machine charm and the prometheus-k8s charm. ### Environment > we have been trying to integrate the prometheus-k8s charm with kube-ovn charm. But it seems that the library gathers the instance-id of our charm instead of the IP address of the unit, therefore metrics are not being collected in the process. > > they are in different models. We have a CK cluster that has the kube-ovn charm as it's CNI charm. > > Kube-OVN is a subordinate machine charm. > -- @mateoflorido ### Relevant log output ```shell ~$ curl 10.152.183.72:9090/api/v1/targets { "status": "success", "data": { "activeTargets": [ { "discoveredLabels": { "__address__": "juju-865c1a-5:10665", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/4", }, "labels": { "instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/4", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/4", }, "scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "scrapeUrl": "http://juju-865c1a-5:10665/metrics", "globalUrl": "http://juju-865c1a-5:10665/metrics", "lastError": 'Get "http://juju-865c1a-5:10665/metrics": dial tcp: lookup juju-865c1a-5 on 10.152.183.213:53: no such host', "lastScrape": "2022-07-15T15:27:59.615508445Z", "lastScrapeDuration": 0.005049553, "health": "down", }, { "discoveredLabels": { "__address__": "juju-865c1a-8:10665", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/0", }, "labels": { "instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/0", "job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "juju_application": "kube-ovn", "juju_charm": "kube-ovn", "juju_model": "test-kovn-prom", "juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a", "juju_unit": "kube-ovn/0", }, "scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape", "scrapeUrl": "http://juju-865c1a-8:10665/metrics", "globalUrl": "http://juju-865c1a-8:10665/metrics", "lastError": 'Get "http://juju-865c1a-8:10665/metrics": dial tcp: lookup juju-865c1a-8 on 10.152.183.213:53: no such host', "lastScrape": "2022-07-15T15:28:30.262574207Z", "lastScrapeDuration": 0.00596523, "health": "down", }, { "discoveredLabels": { "__address__": "localhost:9090", "__metrics_path__": "/metrics", "__scheme__": "http", "job": "prometheus", }, "labels": {"instance": "localhost:9090", "job": "prometheus"}, "scrapePool": "prometheus", "scrapeUrl": "http://localhost:9090/metrics", "globalUrl": "http://192.168.0.17:9090/metrics", "lastError": "", "lastScrape": "2022-07-15T15:28:46.369824673Z", "lastScrapeDuration": 0.006802522, "health": "up", }, ], "droppedTargets": [], }, } ``` ### Additional context _No response_
priority
metricsendpointprovider units are not always reachable via fqdn bug description replaced network bind address with socket getfqdn this works for in model relations but breaks for cross cluster relations cc mateoflorido stonepreston to reproduce form a cross cluster prometheus scrape relation b w a machine charm and the prometheus charm environment we have been trying to integrate the prometheus charm with kube ovn charm but it seems that the library gathers the instance id of our charm instead of the ip address of the unit therefore metrics are not being collected in the process they are in different models we have a ck cluster that has the kube ovn charm as it s cni charm kube ovn is a subordinate machine charm mateoflorido relevant log output shell curl api targets status success data activetargets discoveredlabels address juju metrics path metrics scheme http job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn labels instance test kovn prom acde kube ovn kube ovn job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn scrapepool juju test kovn prom acde kube ovn kube ovn prometheus scrape scrapeurl globalurl lasterror get dial tcp lookup juju on no such host lastscrape lastscrapeduration health down discoveredlabels address juju metrics path metrics scheme http job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn labels instance test kovn prom acde kube ovn kube ovn job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn scrapepool juju test kovn prom acde kube ovn kube ovn prometheus scrape scrapeurl globalurl lasterror get dial tcp lookup juju on no such host lastscrape lastscrapeduration health down discoveredlabels address localhost metrics path metrics scheme http job prometheus labels instance localhost job prometheus scrapepool prometheus scrapeurl globalurl lasterror lastscrape lastscrapeduration health up droppedtargets additional context no response
1
119,700
4,774,618,360
IssuesEvent
2016-10-27 07:31:43
ponylang/ponyc
https://api.github.com/repos/ponylang/ponyc
closed
ponyc built on OS X with llvm-config 3.8 doesn't compile hello world
bug: 1 - needs investigation priority: 3 - high
I pulled master (as of this writing: e4427917a3cccc04dd3633777a31e17b5133a743) and could not build hello world on OS X (El Capitan 10.11.4) when `ponyc` is compiled with LLVM 3.8. I have both LLVM 3.6.2 and LLVM 3.8 installed with homebrew. I'm not sure if anything other than the location of `llvm-config` matters here. Compiling hello world works after compiling pony with LLVM 3.6.2 like so: ``` LLVM_CONFIG=/usr/local/opt/llvm/bin/llvm-config make config=release ``` But it does not work after compiling pony with LLVM 3.8 like so: ``` LLVM_CONFIG=/usr/local/bin/llvm-config-3.8 make config=release ``` With `ponyc` compiled with LLVM 3.8 ``` ponyc:$ cd examples/helloworld/ helloworld:$ ../../build/release/ponyc Building builtin -> /Users/jeremy/src/ponyc/packages/builtin Building . -> /Users/jeremy/src/ponyc/examples/helloworld Generating Reachability Selector painting Data prototypes Data types Function prototypes Descriptors Functions Optimising Writing ./helloworld.o Cannot select: t19: ch = store<ST1[%30]> t16, t7, t18, undef:i64 t7: i1,ch = load<LD1[%23]> t0, t4, undef:i64 t4: i64 = add t2, Constant:i64<16> t2: i64,ch = CopyFromReg t0, Register:i64 %vreg4 t1: i64 = Register %vreg4 t3: i64 = Constant<16> t6: i64 = undef t18: i64 = add t9, Constant:i64<264> t9: i64,ch = CopyFromReg t0, Register:i64 %vreg3 t8: i64 = Register %vreg3 t17: i64 = Constant<264> t6: i64 = undef In function: Stdin_$dispatch ```
1.0
ponyc built on OS X with llvm-config 3.8 doesn't compile hello world - I pulled master (as of this writing: e4427917a3cccc04dd3633777a31e17b5133a743) and could not build hello world on OS X (El Capitan 10.11.4) when `ponyc` is compiled with LLVM 3.8. I have both LLVM 3.6.2 and LLVM 3.8 installed with homebrew. I'm not sure if anything other than the location of `llvm-config` matters here. Compiling hello world works after compiling pony with LLVM 3.6.2 like so: ``` LLVM_CONFIG=/usr/local/opt/llvm/bin/llvm-config make config=release ``` But it does not work after compiling pony with LLVM 3.8 like so: ``` LLVM_CONFIG=/usr/local/bin/llvm-config-3.8 make config=release ``` With `ponyc` compiled with LLVM 3.8 ``` ponyc:$ cd examples/helloworld/ helloworld:$ ../../build/release/ponyc Building builtin -> /Users/jeremy/src/ponyc/packages/builtin Building . -> /Users/jeremy/src/ponyc/examples/helloworld Generating Reachability Selector painting Data prototypes Data types Function prototypes Descriptors Functions Optimising Writing ./helloworld.o Cannot select: t19: ch = store<ST1[%30]> t16, t7, t18, undef:i64 t7: i1,ch = load<LD1[%23]> t0, t4, undef:i64 t4: i64 = add t2, Constant:i64<16> t2: i64,ch = CopyFromReg t0, Register:i64 %vreg4 t1: i64 = Register %vreg4 t3: i64 = Constant<16> t6: i64 = undef t18: i64 = add t9, Constant:i64<264> t9: i64,ch = CopyFromReg t0, Register:i64 %vreg3 t8: i64 = Register %vreg3 t17: i64 = Constant<264> t6: i64 = undef In function: Stdin_$dispatch ```
priority
ponyc built on os x with llvm config doesn t compile hello world i pulled master as of this writing and could not build hello world on os x el capitan when ponyc is compiled with llvm i have both llvm and llvm installed with homebrew i m not sure if anything other than the location of llvm config matters here compiling hello world works after compiling pony with llvm like so llvm config usr local opt llvm bin llvm config make config release but it does not work after compiling pony with llvm like so llvm config usr local bin llvm config make config release with ponyc compiled with llvm ponyc cd examples helloworld helloworld build release ponyc building builtin users jeremy src ponyc packages builtin building users jeremy src ponyc examples helloworld generating reachability selector painting data prototypes data types function prototypes descriptors functions optimising writing helloworld o cannot select ch store undef ch load undef add constant ch copyfromreg register register constant undef add constant ch copyfromreg register register constant undef in function stdin dispatch
1
728,242
25,072,617,266
IssuesEvent
2022-11-07 13:20:08
lazarskiopencourses/lazarskiopencourses.github.io
https://api.github.com/repos/lazarskiopencourses/lazarskiopencourses.github.io
opened
Programming in Python - Advanced Function - Recursion - Fibonacci
bug high priority
There is a mistake in the example code on the website --> `fib(6)` should return `(8, 25)`
1.0
Programming in Python - Advanced Function - Recursion - Fibonacci - There is a mistake in the example code on the website --> `fib(6)` should return `(8, 25)`
priority
programming in python advanced function recursion fibonacci there is a mistake in the example code on the website fib should return
1
761,975
26,705,210,507
IssuesEvent
2023-01-27 17:31:29
evo-lua/evo-runtime
https://api.github.com/repos/evo-lua/evo-runtime
opened
Re-add the event system library so that events can be monitored (for testing purposes)
Priority: High Complexity: Low Scope: Runtime Status: Accepted Type: New Feature
This is the global event registry from evo-luvi (see RFC), but it should be able to buffer events and payloads for testing purposes. Using event emitters makes the code less decoupled and it's a bit of a pain to test whether a server has received some data, for example, which was a problem I encountered when writing tests for the TCP server before. I basically had to write a lot of boilerplate code, when I think it might be easier to just enable the event buffer, do the thing, and then fetch the events and compare to a list (TBD).
1.0
Re-add the event system library so that events can be monitored (for testing purposes) - This is the global event registry from evo-luvi (see RFC), but it should be able to buffer events and payloads for testing purposes. Using event emitters makes the code less decoupled and it's a bit of a pain to test whether a server has received some data, for example, which was a problem I encountered when writing tests for the TCP server before. I basically had to write a lot of boilerplate code, when I think it might be easier to just enable the event buffer, do the thing, and then fetch the events and compare to a list (TBD).
priority
re add the event system library so that events can be monitored for testing purposes this is the global event registry from evo luvi see rfc but it should be able to buffer events and payloads for testing purposes using event emitters makes the code less decoupled and it s a bit of a pain to test whether a server has received some data for example which was a problem i encountered when writing tests for the tcp server before i basically had to write a lot of boilerplate code when i think it might be easier to just enable the event buffer do the thing and then fetch the events and compare to a list tbd
1
283,987
8,729,004,023
IssuesEvent
2018-12-10 19:00:09
conveyal/datatools-ui
https://api.github.com/repos/conveyal/datatools-ui
closed
OTP Deployment: Bounds should be checked for validity before completing deployment
bug high-priority imported
<a href="https://github.com/landonreed"><img src="https://avatars2.githubusercontent.com/u/2370911?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [landonreed](https://github.com/landonreed)** _Friday Jun 01, 2018 at 19:08 GMT_ _Originally opened as https://github.com/catalogueglobal/datatools-ui/issues/189_ ---- A recent deployment job succeeded even though the osm extract was for the bounds `vex/NaN,NaN,NaN,NaN.pbf`. A couple of fixes might be in order: - The server should check that the bounds are valid and fail if not, so that an incomplete set of data is not sent for graph building. - The UI should warn the user that the project bounds are invalid and which feed versions might be problematic.
1.0
OTP Deployment: Bounds should be checked for validity before completing deployment - <a href="https://github.com/landonreed"><img src="https://avatars2.githubusercontent.com/u/2370911?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [landonreed](https://github.com/landonreed)** _Friday Jun 01, 2018 at 19:08 GMT_ _Originally opened as https://github.com/catalogueglobal/datatools-ui/issues/189_ ---- A recent deployment job succeeded even though the osm extract was for the bounds `vex/NaN,NaN,NaN,NaN.pbf`. A couple of fixes might be in order: - The server should check that the bounds are valid and fail if not, so that an incomplete set of data is not sent for graph building. - The UI should warn the user that the project bounds are invalid and which feed versions might be problematic.
priority
otp deployment bounds should be checked for validity before completing deployment issue by friday jun at gmt originally opened as a recent deployment job succeeded even though the osm extract was for the bounds vex nan nan nan nan pbf a couple of fixes might be in order the server should check that the bounds are valid and fail if not so that an incomplete set of data is not sent for graph building the ui should warn the user that the project bounds are invalid and which feed versions might be problematic
1
396,504
11,709,729,755
IssuesEvent
2020-03-08 20:29:56
open-gunz/source
https://api.github.com/repos/open-gunz/source
opened
Server-based netcode and AntiLead
High Priority
Finish the server-based netcode that crawly has made to have real-time discerning between sword hits/rockets/grenades/hitscan weapons/etc. NAT - server sided hit detection. Current Anti-lead issues: Packets sometimes get lost/malformed that can happen once every few 100 packets with UDP. This leads to showing HIT in P2P anti-lead but still does not reduce target's HP/AP. Sometimes, there is no HIT but the target's HP/AP is reduced.
1.0
Server-based netcode and AntiLead - Finish the server-based netcode that crawly has made to have real-time discerning between sword hits/rockets/grenades/hitscan weapons/etc. NAT - server sided hit detection. Current Anti-lead issues: Packets sometimes get lost/malformed that can happen once every few 100 packets with UDP. This leads to showing HIT in P2P anti-lead but still does not reduce target's HP/AP. Sometimes, there is no HIT but the target's HP/AP is reduced.
priority
server based netcode and antilead finish the server based netcode that crawly has made to have real time discerning between sword hits rockets grenades hitscan weapons etc nat server sided hit detection current anti lead issues packets sometimes get lost malformed that can happen once every few packets with udp this leads to showing hit in anti lead but still does not reduce target s hp ap sometimes there is no hit but the target s hp ap is reduced
1
517,138
14,995,403,885
IssuesEvent
2021-01-29 14:16:39
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
Photo more options hover, photo popup shows previous image for a while
bug priority: high
**Describe the bug** 1. Photos (...) ellipsis dropdown hover is not proper 2. Photos Popup behaving weird by showing last open image and then new select image. 3. View Group Photos in Popup as Organizer > Fix order here to show “Report” just before “delete” button: https://prnt.sc/xpffcs **Video** https://www.loom.com/share/8a9e51400ce2466387ce520f31be768d
1.0
Photo more options hover, photo popup shows previous image for a while - **Describe the bug** 1. Photos (...) ellipsis dropdown hover is not proper 2. Photos Popup behaving weird by showing last open image and then new select image. 3. View Group Photos in Popup as Organizer > Fix order here to show “Report” just before “delete” button: https://prnt.sc/xpffcs **Video** https://www.loom.com/share/8a9e51400ce2466387ce520f31be768d
priority
photo more options hover photo popup shows previous image for a while describe the bug photos ellipsis dropdown hover is not proper photos popup behaving weird by showing last open image and then new select image view group photos in popup as organizer fix order here to show “report” just before “delete” button video
1
89,289
3,792,076,257
IssuesEvent
2016-03-22 07:49:16
cs2103jan2016-W10-1J/main
https://api.github.com/repos/cs2103jan2016-W10-1J/main
closed
Logic support according to Architectural change adopting inheritance and polymorphism
priority.high
- [x] Code for Update Class as part of the Logic - [x] Code for the Add Class as part of the Logic
1.0
Logic support according to Architectural change adopting inheritance and polymorphism - - [x] Code for Update Class as part of the Logic - [x] Code for the Add Class as part of the Logic
priority
logic support according to architectural change adopting inheritance and polymorphism code for update class as part of the logic code for the add class as part of the logic
1
172,258
6,500,824,159
IssuesEvent
2017-08-23 07:02:01
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
opened
[xmlLiterals] Giving syntax errors as \ will be removed
Bug Priority:High Severity:Major
Pack - 23/08/2017 This given as error as the slash (\) will be removed when you switch to the sources view adding this from the design view `xml x5 = xml `aaa{{v1}}b\{{bb{{v2}}c\}}cc{d{}e}{f{`;`
1.0
[xmlLiterals] Giving syntax errors as \ will be removed - Pack - 23/08/2017 This given as error as the slash (\) will be removed when you switch to the sources view adding this from the design view `xml x5 = xml `aaa{{v1}}b\{{bb{{v2}}c\}}cc{d{}e}{f{`;`
priority
giving syntax errors as will be removed pack this given as error as the slash will be removed when you switch to the sources view adding this from the design view xml xml aaa b bb c cc d e f
1
439,159
12,678,386,926
IssuesEvent
2020-06-19 09:40:50
localstack/localstack
https://api.github.com/repos/localstack/localstack
closed
Unable to determine forwarding port for API "monitoring"
bug priority-high
<!-- Love localstack? Please consider supporting our collective: 👉 https://opencollective.com/localstack/donate --> # Type of request: This is a ... [x ] bug report [ ] feature request # Detailed description Not able to use the edge service to write metric data to cloudwatch via the javascript aws-sdk. However, It **DOES** work when using the deprecated cloudwatch port. ## Expected behavior Metric data to be visible when calling `awslocal cloudwatch list-metrics` ## Actual behavior 404 from edge service. No metric data is persisted, localstack instance logs the following info: ``` INFO:localstack.services.edge: Unable to determine forwarding port for API "monitoring" - please make sure this API is enabled via the SERVICES configuration ``` # Steps to reproduce Run JS code in "Client code" section. See 404 Error ## Command used to start LocalStack `TMPDIR=/tmp/localstack docker-compose up` ## Client code (AWS SDK code snippet, or sequence of "awslocal" commands) ```js const AWS = require('aws-sdk') const cw = new AWS.CloudWatch({ endpoint: 'http://localhost:4566', }) const data = { Namespace: 'ns', MetricData: [ { MetricName: 'metric-name', Timestamp: new Date(), Value: 12345, Unit: 'Milliseconds', Dimensions: [ { Name: 'Environment', Value: 'local' }, ] } ] } cw.putMetricData(data, (err, data) => { if (err) { console.log('Error', err) } else { console.log('Success', JSON.stringify(data)) } }) ```
1.0
Unable to determine forwarding port for API "monitoring" - <!-- Love localstack? Please consider supporting our collective: 👉 https://opencollective.com/localstack/donate --> # Type of request: This is a ... [x ] bug report [ ] feature request # Detailed description Not able to use the edge service to write metric data to cloudwatch via the javascript aws-sdk. However, It **DOES** work when using the deprecated cloudwatch port. ## Expected behavior Metric data to be visible when calling `awslocal cloudwatch list-metrics` ## Actual behavior 404 from edge service. No metric data is persisted, localstack instance logs the following info: ``` INFO:localstack.services.edge: Unable to determine forwarding port for API "monitoring" - please make sure this API is enabled via the SERVICES configuration ``` # Steps to reproduce Run JS code in "Client code" section. See 404 Error ## Command used to start LocalStack `TMPDIR=/tmp/localstack docker-compose up` ## Client code (AWS SDK code snippet, or sequence of "awslocal" commands) ```js const AWS = require('aws-sdk') const cw = new AWS.CloudWatch({ endpoint: 'http://localhost:4566', }) const data = { Namespace: 'ns', MetricData: [ { MetricName: 'metric-name', Timestamp: new Date(), Value: 12345, Unit: 'Milliseconds', Dimensions: [ { Name: 'Environment', Value: 'local' }, ] } ] } cw.putMetricData(data, (err, data) => { if (err) { console.log('Error', err) } else { console.log('Success', JSON.stringify(data)) } }) ```
priority
unable to determine forwarding port for api monitoring love localstack please consider supporting our collective 👉 type of request this is a bug report feature request detailed description not able to use the edge service to write metric data to cloudwatch via the javascript aws sdk however it does work when using the deprecated cloudwatch port expected behavior metric data to be visible when calling awslocal cloudwatch list metrics actual behavior from edge service no metric data is persisted localstack instance logs the following info info localstack services edge unable to determine forwarding port for api monitoring please make sure this api is enabled via the services configuration steps to reproduce run js code in client code section see error command used to start localstack tmpdir tmp localstack docker compose up client code aws sdk code snippet or sequence of awslocal commands js const aws require aws sdk const cw new aws cloudwatch endpoint const data namespace ns metricdata metricname metric name timestamp new date value unit milliseconds dimensions name environment value local cw putmetricdata data err data if err console log error err else console log success json stringify data
1
439,359
12,681,723,483
IssuesEvent
2020-06-19 15:55:20
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
opened
Task stats in User DTO
Component: Backend Difficulty: 3 Priority: High Type: Bug
`/users/queries/username` endpoint is missing details on task stats - mapped/validated/invalidated
1.0
Task stats in User DTO - `/users/queries/username` endpoint is missing details on task stats - mapped/validated/invalidated
priority
task stats in user dto users queries username endpoint is missing details on task stats mapped validated invalidated
1
281,402
8,694,755,567
IssuesEvent
2018-12-04 13:33:13
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
toFQDNs with no IP data do not block traffic
area/daemon area/policy kind/bug priority/high
When a toFQDN rule has L4/L7 sections, it will incorrectly interpret the L3 as a wildcard allow-all. This means it isn't really enforcing what one expects. When IPs are inserted into toCIDRSet the enforcement is correct. This policy works incorrectly with our minikube GSG: ``` apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: rebel-escape namespace: default spec: endpointSelector: matchLabels: any:org: alliance egress: - toEndpoints: - matchLabels: "k8s:io.kubernetes.pod.namespace": kube-system "k8s:k8s-app": kube-dns - toFQDNs: - matchName: "notasubdomain.cilium.io" toPorts: - ports: - port: "80" ``` Both of these work when they should not (the port is enforced, as would a L7 HTTP rules) `kubectl exec xwing -- curl -s cilium.io` `kubectl exec xwing -- curl -s google.com`
1.0
toFQDNs with no IP data do not block traffic - When a toFQDN rule has L4/L7 sections, it will incorrectly interpret the L3 as a wildcard allow-all. This means it isn't really enforcing what one expects. When IPs are inserted into toCIDRSet the enforcement is correct. This policy works incorrectly with our minikube GSG: ``` apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: rebel-escape namespace: default spec: endpointSelector: matchLabels: any:org: alliance egress: - toEndpoints: - matchLabels: "k8s:io.kubernetes.pod.namespace": kube-system "k8s:k8s-app": kube-dns - toFQDNs: - matchName: "notasubdomain.cilium.io" toPorts: - ports: - port: "80" ``` Both of these work when they should not (the port is enforced, as would a L7 HTTP rules) `kubectl exec xwing -- curl -s cilium.io` `kubectl exec xwing -- curl -s google.com`
priority
tofqdns with no ip data do not block traffic when a tofqdn rule has sections it will incorrectly interpret the as a wildcard allow all this means it isn t really enforcing what one expects when ips are inserted into tocidrset the enforcement is correct this policy works incorrectly with our minikube gsg apiversion cilium io kind ciliumnetworkpolicy metadata name rebel escape namespace default spec endpointselector matchlabels any org alliance egress toendpoints matchlabels io kubernetes pod namespace kube system app kube dns tofqdns matchname notasubdomain cilium io toports ports port both of these work when they should not the port is enforced as would a http rules kubectl exec xwing curl s cilium io kubectl exec xwing curl s google com
1
805,460
29,520,516,223
IssuesEvent
2023-06-05 01:00:48
steedos/steedos-platform
https://api.github.com/repos/steedos/steedos-platform
closed
[Bug]: 审批,待审核/监控箱等列表,搜素,流程字段统一为新建流程的式样
bug done priority: High
### Description 目前是普通lookup类型字段的式样 ![image](https://github.com/steedos/steedos-platform/assets/7553007/c04dbb72-5771-4389-9018-14de0101f63b) 应统一为 新建流程时的选项式样 ![image](https://github.com/steedos/steedos-platform/assets/7553007/cebaac7d-7af5-44a5-9755-538da8da74cf) BTW,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些 ### Steps To Reproduce 重现步骤 审批,待审核/监控箱等列表,搜素,流程 ### Version 版本 2.5.0-beta.39
1.0
[Bug]: 审批,待审核/监控箱等列表,搜素,流程字段统一为新建流程的式样 - ### Description 目前是普通lookup类型字段的式样 ![image](https://github.com/steedos/steedos-platform/assets/7553007/c04dbb72-5771-4389-9018-14de0101f63b) 应统一为 新建流程时的选项式样 ![image](https://github.com/steedos/steedos-platform/assets/7553007/cebaac7d-7af5-44a5-9755-538da8da74cf) BTW,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些 ### Steps To Reproduce 重现步骤 审批,待审核/监控箱等列表,搜素,流程 ### Version 版本 2.5.0-beta.39
priority
审批,待审核 监控箱等列表,搜素,流程字段统一为新建流程的式样 description 目前是普通lookup类型字段的式样 应统一为 新建流程时的选项式样 btw,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些 steps to reproduce 重现步骤 审批,待审核 监控箱等列表,搜素,流程 version 版本 beta
1
288,869
8,852,522,493
IssuesEvent
2019-01-08 18:36:14
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
VisIt crashes on startup on Windows Vista
bug crash likelihood high priority reviewed severity high wrong results
When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 192 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: VisIt crashes on startup on Windows Vista Assigned to: Kathleen Biagas Category: - Target version: 2.1 Author: Kathleen Biagas Start: 06/30/2010 Due date: % Done: 0% Estimated time: Created: 06/30/2010 08:19 pm Updated: 08/27/2010 05:34 pm Likelihood: 4 - Common Severity: 4 - Crash / Wrong Results Found in version: 2.0.0 Impact: Expected Use: OS: Windows Support Group: Any Description: When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2. Comments: Assignment from LLNL VisIt 2.1 Release Meeting I downloaded Microsoft's 'Application Compatibility Toolkit' in order to track down why Vista is flagging VisIt as needing to be run in compatibility mode (especially an NT4 compat mode!)First few passes with the tool appear to indicate that VisIt is attempting to WRITE to HKLM registry files, which should notbe happening. That's the only compatibility issues that cropped up. The registry write does not appear in VisIt's source code.I ran VisIt through a a tool that generates callstack information to discover the source of the Registry write operations, andit appears to be happening down in GL calls. (wglSwapMultipleBuffers). If this is truly the case, I'm not sure what we can doto mitigate this issue on Vista. Binaries built with Visual Studio 9.0 (2008) do not have the same issue.Also, binaries built on Vista using Visual Studio 8 do not have this issue.Starting with VisIt 2.1, we will distribute binaries built with Visual Studio 9.
1.0
VisIt crashes on startup on Windows Vista - When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 192 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: VisIt crashes on startup on Windows Vista Assigned to: Kathleen Biagas Category: - Target version: 2.1 Author: Kathleen Biagas Start: 06/30/2010 Due date: % Done: 0% Estimated time: Created: 06/30/2010 08:19 pm Updated: 08/27/2010 05:34 pm Likelihood: 4 - Common Severity: 4 - Crash / Wrong Results Found in version: 2.0.0 Impact: Expected Use: OS: Windows Support Group: Any Description: When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2. Comments: Assignment from LLNL VisIt 2.1 Release Meeting I downloaded Microsoft's 'Application Compatibility Toolkit' in order to track down why Vista is flagging VisIt as needing to be run in compatibility mode (especially an NT4 compat mode!)First few passes with the tool appear to indicate that VisIt is attempting to WRITE to HKLM registry files, which should notbe happening. That's the only compatibility issues that cropped up. The registry write does not appear in VisIt's source code.I ran VisIt through a a tool that generates callstack information to discover the source of the Registry write operations, andit appears to be happening down in GL calls. (wglSwapMultipleBuffers). If this is truly the case, I'm not sure what we can doto mitigate this issue on Vista. Binaries built with Visual Studio 9.0 (2008) do not have the same issue.Also, binaries built on Vista using Visual Studio 8 do not have this issue.Starting with VisIt 2.1, we will distribute binaries built with Visual Studio 9.
priority
visit crashes on startup on windows vista when starting visit on windows vista the mdserver crashes on startup a few users have reported this workaround to run visit with compatibility mode set to nt service pack i verified this work around on my version of vista which is vista business bit redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit crashes on startup on windows vista assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood common severity crash wrong results found in version impact expected use os windows support group any description when starting visit on windows vista the mdserver crashes on startup a few users have reported this workaround to run visit with compatibility mode set to nt service pack i verified this work around on my version of vista which is vista business bit comments assignment from llnl visit release meeting i downloaded microsoft s application compatibility toolkit in order to track down why vista is flagging visit as needing to be run in compatibility mode especially an compat mode first few passes with the tool appear to indicate that visit is attempting to write to hklm registry files which should notbe happening that s the only compatibility issues that cropped up the registry write does not appear in visit s source code i ran visit through a a tool that generates callstack information to discover the source of the registry write operations andit appears to be happening down in gl calls wglswapmultiplebuffers if this is truly the case i m not sure what we can doto mitigate this issue on vista binaries built with visual studio do not have the same issue also binaries built on vista using visual studio do not have this issue starting with visit we will distribute binaries built with visual studio
1
815,678
30,567,189,246
IssuesEvent
2023-07-20 18:44:21
DiscoTrayStudios/hendrix-today
https://api.github.com/repos/DiscoTrayStudios/hendrix-today
closed
Disco Tray Logo and credits
category: ui priority: high size: small type: enhancement
There should be a mention somewhere of Disco Tray Studios with our logo and a link to the website, along with the names of the student developers who worked on the app.
1.0
Disco Tray Logo and credits - There should be a mention somewhere of Disco Tray Studios with our logo and a link to the website, along with the names of the student developers who worked on the app.
priority
disco tray logo and credits there should be a mention somewhere of disco tray studios with our logo and a link to the website along with the names of the student developers who worked on the app
1
711,897
24,479,110,831
IssuesEvent
2022-10-08 15:27:09
JasonBock/Rocks
https://api.github.com/repos/JasonBock/Rocks
closed
Duplicating Overrides From `new` Methods
bug High Priority
```csharp #nullable enable public class PropertyBuilder { public virtual PropertyBuilder HasAnnotation(string annotation, object? value) => default!; } public class PropertyBuilder<TProperty> : PropertyBuilder { public new virtual PropertyBuilder<TProperty> HasAnnotation(string annotation, object? value) => default!; } public static class Test { public static void Go() { var expectations = Rock.Create<PropertyBuilder<object>>(); } } ``` What the mock type should generate is just the one method from the generic class, not both of them. So I'm guessing I need logic to see if a method is "shadowing" a base one via `new`.
1.0
Duplicating Overrides From `new` Methods - ```csharp #nullable enable public class PropertyBuilder { public virtual PropertyBuilder HasAnnotation(string annotation, object? value) => default!; } public class PropertyBuilder<TProperty> : PropertyBuilder { public new virtual PropertyBuilder<TProperty> HasAnnotation(string annotation, object? value) => default!; } public static class Test { public static void Go() { var expectations = Rock.Create<PropertyBuilder<object>>(); } } ``` What the mock type should generate is just the one method from the generic class, not both of them. So I'm guessing I need logic to see if a method is "shadowing" a base one via `new`.
priority
duplicating overrides from new methods csharp nullable enable public class propertybuilder public virtual propertybuilder hasannotation string annotation object value default public class propertybuilder propertybuilder public new virtual propertybuilder hasannotation string annotation object value default public static class test public static void go var expectations rock create what the mock type should generate is just the one method from the generic class not both of them so i m guessing i need logic to see if a method is shadowing a base one via new
1
707,659
24,313,213,472
IssuesEvent
2022-09-30 02:02:31
AY2223S1-CS2103T-T15-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-T15-1/tp
closed
Update About Us Document
priority.high type.about-us
The "About Us" document/page should contain details about the team working on the project (that's us), and we should reflect that. The document should contain a simple profile for each of us, which should include: - A profile picture (pick a nice one) - Your name / github username - Your roles & responsibilities (as we have discussed) - Anything else you want to add for fun With this, the subtasks involved are for each person to update and add in their own details: - Chao Yung #19 - Kevin #10 - Prittam #20 - Shawn #21 - Shi Jun #22
1.0
Update About Us Document - The "About Us" document/page should contain details about the team working on the project (that's us), and we should reflect that. The document should contain a simple profile for each of us, which should include: - A profile picture (pick a nice one) - Your name / github username - Your roles & responsibilities (as we have discussed) - Anything else you want to add for fun With this, the subtasks involved are for each person to update and add in their own details: - Chao Yung #19 - Kevin #10 - Prittam #20 - Shawn #21 - Shi Jun #22
priority
update about us document the about us document page should contain details about the team working on the project that s us and we should reflect that the document should contain a simple profile for each of us which should include a profile picture pick a nice one your name github username your roles responsibilities as we have discussed anything else you want to add for fun with this the subtasks involved are for each person to update and add in their own details chao yung kevin prittam shawn shi jun
1
419,147
12,218,234,580
IssuesEvent
2020-05-01 18:53:58
nextstrain/auspice
https://api.github.com/repos/nextstrain/auspice
opened
Provide toggle to display transmission lines in the map
enhancement high priority moderate problem please take this issue
**Context** With SARS-CoV-2 jumping around the globe so quickly, showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding. See for example: <img width="1134" alt="lines" src="https://user-images.githubusercontent.com/1176109/80832435-807ea380-8ba1-11ea-9418-b0972f8c21f3.png"> There are frequency situations where we want "traits" reconstruction on phylogeny, but don't want to show these lines on the map. **Description** Include a toggle in the sidebar under "Map Options" directly below "Geographic Resolution" that says "Show transmission lines". When toggled on, we'd have the current behavior. When toggled off, the deme circles / pie charts would remain but no transmission lines would be drawn on the map. In addition, this toggle should update the URL and (eventually) this "show transmission lines" default behavior should be included in the Auspice JSON as something like: ``` "display_defaults": { "transmission_lines": true }, ``` But would be good to get working as UI element and URL before bothering to include in JSON.
1.0
Provide toggle to display transmission lines in the map - **Context** With SARS-CoV-2 jumping around the globe so quickly, showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding. See for example: <img width="1134" alt="lines" src="https://user-images.githubusercontent.com/1176109/80832435-807ea380-8ba1-11ea-9418-b0972f8c21f3.png"> There are frequency situations where we want "traits" reconstruction on phylogeny, but don't want to show these lines on the map. **Description** Include a toggle in the sidebar under "Map Options" directly below "Geographic Resolution" that says "Show transmission lines". When toggled on, we'd have the current behavior. When toggled off, the deme circles / pie charts would remain but no transmission lines would be drawn on the map. In addition, this toggle should update the URL and (eventually) this "show transmission lines" default behavior should be included in the Auspice JSON as something like: ``` "display_defaults": { "transmission_lines": true }, ``` But would be good to get working as UI element and URL before bothering to include in JSON.
priority
provide toggle to display transmission lines in the map context with sars cov jumping around the globe so quickly showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding see for example img width alt lines src there are frequency situations where we want traits reconstruction on phylogeny but don t want to show these lines on the map description include a toggle in the sidebar under map options directly below geographic resolution that says show transmission lines when toggled on we d have the current behavior when toggled off the deme circles pie charts would remain but no transmission lines would be drawn on the map in addition this toggle should update the url and eventually this show transmission lines default behavior should be included in the auspice json as something like display defaults transmission lines true but would be good to get working as ui element and url before bothering to include in json
1
236,838
7,753,052,868
IssuesEvent
2018-05-30 22:33:05
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0004294: renaming a resource crashes
Calendar bug high priority
**Reported by pschuele on 13 May 2011 09:24** **Version:** Maischa (2011-05) Beta 1 renaming a resource crashes **Additional information:** $_record is of invalid model type .../Calendar/Controller/Resource.php(122): Tinebase_Backend_Sql_Abstract-&gt;update() [internal function]: Calendar_Controller_Resource-&gt;update() .../Tinebase/Frontend/Json/Abstract.php(163): call_user_func_array() .../Calendar/Frontend/Json.php(200): Tinebase_Frontend_Json_Abstract-&gt;_save() [internal function]: Calendar_Frontend_Json-&gt;saveResource() .../library/Zend/Server/Abstract.php(232): call_user_func_array() .../Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch() .../Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() .../Tinebase/Server/Json.php(140): Zend_Json_Server-&gt;handle() .../Tinebase/Server/Json.php(76): Tinebase_Server_Json-&gt;_handle() .../Tinebase/Core.php(223): Tinebase_Server_Json-&gt;handle() .../index.php(57): Tinebase_Core::dispatchRequest()
1.0
0004294: renaming a resource crashes - **Reported by pschuele on 13 May 2011 09:24** **Version:** Maischa (2011-05) Beta 1 renaming a resource crashes **Additional information:** $_record is of invalid model type .../Calendar/Controller/Resource.php(122): Tinebase_Backend_Sql_Abstract-&gt;update() [internal function]: Calendar_Controller_Resource-&gt;update() .../Tinebase/Frontend/Json/Abstract.php(163): call_user_func_array() .../Calendar/Frontend/Json.php(200): Tinebase_Frontend_Json_Abstract-&gt;_save() [internal function]: Calendar_Frontend_Json-&gt;saveResource() .../library/Zend/Server/Abstract.php(232): call_user_func_array() .../Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch() .../Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() .../Tinebase/Server/Json.php(140): Zend_Json_Server-&gt;handle() .../Tinebase/Server/Json.php(76): Tinebase_Server_Json-&gt;_handle() .../Tinebase/Core.php(223): Tinebase_Server_Json-&gt;handle() .../index.php(57): Tinebase_Core::dispatchRequest()
priority
renaming a resource crashes reported by pschuele on may version maischa beta renaming a resource crashes additional information record is of invalid model type calendar controller resource php tinebase backend sql abstract gt update calendar controller resource gt update tinebase frontend json abstract php call user func array calendar frontend json php tinebase frontend json abstract gt save calendar frontend json gt saveresource library zend server abstract php call user func array zend json server php zend server abstract gt dispatch zend json server php zend json server gt handle tinebase server json php zend json server gt handle tinebase server json php tinebase server json gt handle tinebase core php tinebase server json gt handle index php tinebase core dispatchrequest
1
31,120
2,732,045,414
IssuesEvent
2015-04-17 00:52:05
EFForg/privacybadgerchrome
https://api.github.com/repos/EFForg/privacybadgerchrome
opened
Memory leak
bug High priority
We have a memory leak, in my browser privacy badger is using about 1.5G of memory.
1.0
Memory leak - We have a memory leak, in my browser privacy badger is using about 1.5G of memory.
priority
memory leak we have a memory leak in my browser privacy badger is using about of memory
1
348,804
10,452,991,241
IssuesEvent
2019-09-19 15:40:29
openshift/odo
https://api.github.com/repos/openshift/odo
closed
`odo component list -o json` returns empty list for `not pushed` components
kind/bug priority/High
[kind/bug] `odo component list` should display even components that are not pushed. The same problem is with `odo component describe` It should show information even for components that are not pushed to the cluster. <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the Google group if you have a question rather than a bug or feature request. The group is at: https://groups.google.com/forum/#!forum/odo-users Thanks for understanding, and for contributing to the project! --> ## How did you run odo exactly? `odo component list -o json | jq .` ## Actual behavior ``` ▶ odo component list -o json | jq . { "kind": "List", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": {}, "items": [] } ``` ``` ▶ odo component list ✗ There are no components deployed. ``` ## Expected behavior ``` ▶ odo component list -o json | jq . { "kind": "List", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": {}, "items": [ { "kind": "Component", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": { "name": "java-spring-boot-addx", "namespace": "asdf", "creationTimestamp": null }, "spec": { "app": "app", "type": "java", "source": "file://./", "url": [ "java-spring-boot-addx-8080" ] }, "status": { "state": "Not pushed" } } ] } ``` ``` ▶ odo component list APP NAME TYPE SOURCE STATE app java-spring-boot-addx java file://./ Not pushed ```
1.0
`odo component list -o json` returns empty list for `not pushed` components - [kind/bug] `odo component list` should display even components that are not pushed. The same problem is with `odo component describe` It should show information even for components that are not pushed to the cluster. <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the Google group if you have a question rather than a bug or feature request. The group is at: https://groups.google.com/forum/#!forum/odo-users Thanks for understanding, and for contributing to the project! --> ## How did you run odo exactly? `odo component list -o json | jq .` ## Actual behavior ``` ▶ odo component list -o json | jq . { "kind": "List", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": {}, "items": [] } ``` ``` ▶ odo component list ✗ There are no components deployed. ``` ## Expected behavior ``` ▶ odo component list -o json | jq . { "kind": "List", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": {}, "items": [ { "kind": "Component", "apiVersion": "odo.openshift.io/v1alpha1", "metadata": { "name": "java-spring-boot-addx", "namespace": "asdf", "creationTimestamp": null }, "spec": { "app": "app", "type": "java", "source": "file://./", "url": [ "java-spring-boot-addx-8080" ] }, "status": { "state": "Not pushed" } } ] } ``` ``` ▶ odo component list APP NAME TYPE SOURCE STATE app java-spring-boot-addx java file://./ Not pushed ```
priority
odo component list o json returns empty list for not pushed components odo component list should display even components that are not pushed the same problem is with odo component describe it should show information even for components that are not pushed to the cluster welcome we kindly ask you to fill out the issue template below use the google group if you have a question rather than a bug or feature request the group is at thanks for understanding and for contributing to the project how did you run odo exactly odo component list o json jq actual behavior ▶ odo component list o json jq kind list apiversion odo openshift io metadata items ▶ odo component list ✗ there are no components deployed expected behavior ▶ odo component list o json jq kind list apiversion odo openshift io metadata items kind component apiversion odo openshift io metadata name java spring boot addx namespace asdf creationtimestamp null spec app app type java source file url java spring boot addx status state not pushed ▶ odo component list app name type source state app java spring boot addx java file not pushed
1
685,779
23,467,034,792
IssuesEvent
2022-08-16 17:47:40
larsiusprime/tdrpg-bugs
https://api.github.com/repos/larsiusprime/tdrpg-bugs
closed
The Markos/Ketta Bad Ending still plays 1 second of the sad song, then loads the normal ending music.
bug DQ CORE Cutscene 1 Please Verify Priority HIGH
I thought this had been fixed a few times in DQold. How does this keep popping up? :P
1.0
The Markos/Ketta Bad Ending still plays 1 second of the sad song, then loads the normal ending music. - I thought this had been fixed a few times in DQold. How does this keep popping up? :P
priority
the markos ketta bad ending still plays second of the sad song then loads the normal ending music i thought this had been fixed a few times in dqold how does this keep popping up p
1
693,553
23,781,237,203
IssuesEvent
2022-09-02 05:12:25
younginnovations/iatipublisher
https://api.github.com/repos/younginnovations/iatipublisher
closed
Bug :Activity Detail>>Publishing issue
type: bug priority: high Backend
Context - Desktop - Chrome 102.0.5005.61 Precondition - https://stage.iatipublisher.yipl.com.np/ - Username: Publisher 3 - Password: test1234 - for created activity - [x] **Issue 1 :After publishing and editing activity data, the user cannot unpublish the activity data.** steps 1. publish activity data 2. After publishing activity data edit the activity data Actual Result ![Screenshot from 2022-08-23 14-38-14](https://user-images.githubusercontent.com/78422663/186116096-ac169ba2-a932-4a24-a9b5-725eded25757.png) Excepted Result ![Screenshot from 2022-08-23 14-41-16](https://user-images.githubusercontent.com/78422663/186116696-a6ea4af1-0474-4e6d-b2c2-6dce41f311ff.png) - [x] **Issue 2 : Error message is displayed after unpublishing the data.** Actual Result ![Screenshot from 2022-08-23 14-48-07](https://user-images.githubusercontent.com/78422663/186118186-8dad4c7d-1823-417b-a3dd-f1e49441d9f3.png) Excepted Result - Error message should not displayed after unpublishing the data
1.0
Bug :Activity Detail>>Publishing issue - Context - Desktop - Chrome 102.0.5005.61 Precondition - https://stage.iatipublisher.yipl.com.np/ - Username: Publisher 3 - Password: test1234 - for created activity - [x] **Issue 1 :After publishing and editing activity data, the user cannot unpublish the activity data.** steps 1. publish activity data 2. After publishing activity data edit the activity data Actual Result ![Screenshot from 2022-08-23 14-38-14](https://user-images.githubusercontent.com/78422663/186116096-ac169ba2-a932-4a24-a9b5-725eded25757.png) Excepted Result ![Screenshot from 2022-08-23 14-41-16](https://user-images.githubusercontent.com/78422663/186116696-a6ea4af1-0474-4e6d-b2c2-6dce41f311ff.png) - [x] **Issue 2 : Error message is displayed after unpublishing the data.** Actual Result ![Screenshot from 2022-08-23 14-48-07](https://user-images.githubusercontent.com/78422663/186118186-8dad4c7d-1823-417b-a3dd-f1e49441d9f3.png) Excepted Result - Error message should not displayed after unpublishing the data
priority
bug activity detail publishing issue context desktop chrome precondition username publisher password for created activity issue after publishing and editing activity data the user cannot unpublish the activity data steps publish activity data after publishing activity data edit the activity data actual result excepted result issue error message is displayed after unpublishing the data actual result excepted result error message should not displayed after unpublishing the data
1
806,863
29,923,872,473
IssuesEvent
2023-06-22 02:28:57
TEAM-cafe-in/cafe-in-be
https://api.github.com/repos/TEAM-cafe-in/cafe-in-be
closed
feature: 카페 상세정보 조회 API를 구현한다
🔥High priority
### As-is --- - - ### To-be - [x] MemberInfoController response에 ViewedCafe 리스트를 반환하도록 변경한다 - [x] 커피콩 차감 로직을 구현한다. 카페 조회 시 조회한 카페 목록을 추가한다 - [x] CafeInfo API를 구현한다
1.0
feature: 카페 상세정보 조회 API를 구현한다 - ### As-is --- - - ### To-be - [x] MemberInfoController response에 ViewedCafe 리스트를 반환하도록 변경한다 - [x] 커피콩 차감 로직을 구현한다. 카페 조회 시 조회한 카페 목록을 추가한다 - [x] CafeInfo API를 구현한다
priority
feature 카페 상세정보 조회 api를 구현한다 as is to be memberinfocontroller response에 viewedcafe 리스트를 반환하도록 변경한다 커피콩 차감 로직을 구현한다 카페 조회 시 조회한 카페 목록을 추가한다 cafeinfo api를 구현한다
1
176,795
6,565,305,835
IssuesEvent
2017-09-08 07:44:30
OpenEMS/openems
https://api.github.com/repos/OpenEMS/openems
closed
refactor monitoring
Component: UI Priority: High Type: Enhancement
<!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING --> ### Bug Report or Feature Request (mark with an `x`) ``` - [ ] bug report -> please search issues before submitting - [x ] feature request ``` ### Bug description or desired functionality. <!-- What would like to see implemented? What is the usecase? --> The Online-Monitoring should run more stable and faster
1.0
refactor monitoring - <!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING --> ### Bug Report or Feature Request (mark with an `x`) ``` - [ ] bug report -> please search issues before submitting - [x ] feature request ``` ### Bug description or desired functionality. <!-- What would like to see implemented? What is the usecase? --> The Online-Monitoring should run more stable and faster
priority
refactor monitoring if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request bug description or desired functionality what would like to see implemented what is the usecase the online monitoring should run more stable and faster
1
39,577
2,856,985,822
IssuesEvent
2015-06-02 17:30:16
ChrisMahlke/contribute-dev
https://api.github.com/repos/ChrisMahlke/contribute-dev
closed
Curator: once an item is in Nomcur, if it goes unshared, warn the curator
HighPriority
Example: Rupert nominated his storymap. Deane then accepted it on production. Rupert made a change, such that the storymap's item is no longer available from http://www.arcgis.com/home/item.html?id=9c1100e474e44c55bf9edbe49e35fa19 The record is still visible in Nomcur table, but the app does not warn the curator. This is low-impact, since the user knows he unshared it, but to the curator, it's as if this item never existed. A possible solution: have an option to view records like this as a filter.
1.0
Curator: once an item is in Nomcur, if it goes unshared, warn the curator - Example: Rupert nominated his storymap. Deane then accepted it on production. Rupert made a change, such that the storymap's item is no longer available from http://www.arcgis.com/home/item.html?id=9c1100e474e44c55bf9edbe49e35fa19 The record is still visible in Nomcur table, but the app does not warn the curator. This is low-impact, since the user knows he unshared it, but to the curator, it's as if this item never existed. A possible solution: have an option to view records like this as a filter.
priority
curator once an item is in nomcur if it goes unshared warn the curator example rupert nominated his storymap deane then accepted it on production rupert made a change such that the storymap s item is no longer available from the record is still visible in nomcur table but the app does not warn the curator this is low impact since the user knows he unshared it but to the curator it s as if this item never existed a possible solution have an option to view records like this as a filter
1
335,144
10,149,370,454
IssuesEvent
2019-08-05 15:04:55
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Investigate Accuracy regression on MobileNetV2
high priority module: vision triaged
Duplicate of https://github.com/pytorch/vision/issues/1172, but in main torch repo so we can keep all milestone issues together.
1.0
Investigate Accuracy regression on MobileNetV2 - Duplicate of https://github.com/pytorch/vision/issues/1172, but in main torch repo so we can keep all milestone issues together.
priority
investigate accuracy regression on duplicate of but in main torch repo so we can keep all milestone issues together
1
52,205
3,022,259,491
IssuesEvent
2015-07-31 19:14:41
EFForg/privacybadgerfirefox
https://api.github.com/repos/EFForg/privacybadgerfirefox
closed
Top level domain user settings override subdomain user settings
bug high-priority
If I set an individual subdomain to red and then the top level domain to yellow, I see that the individual subdomain gets put under the etld as cookie blocked, it should be blocked. Similarly if I have a top level domain that is set to green, any subdomains which I previously had set to yellow or red get put under the green top level domain. So it seems like etld/green trumps subdomain red and yellow, etld yellow trumps subdomain red and etld red trumps no subdomain. I feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it's top level domain. (An alternative argument could be made that it should always be grouped under it's parent domain, either way it should be consistent.)
1.0
Top level domain user settings override subdomain user settings - If I set an individual subdomain to red and then the top level domain to yellow, I see that the individual subdomain gets put under the etld as cookie blocked, it should be blocked. Similarly if I have a top level domain that is set to green, any subdomains which I previously had set to yellow or red get put under the green top level domain. So it seems like etld/green trumps subdomain red and yellow, etld yellow trumps subdomain red and etld red trumps no subdomain. I feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it's top level domain. (An alternative argument could be made that it should always be grouped under it's parent domain, either way it should be consistent.)
priority
top level domain user settings override subdomain user settings if i set an individual subdomain to red and then the top level domain to yellow i see that the individual subdomain gets put under the etld as cookie blocked it should be blocked similarly if i have a top level domain that is set to green any subdomains which i previously had set to yellow or red get put under the green top level domain so it seems like etld green trumps subdomain red and yellow etld yellow trumps subdomain red and etld red trumps no subdomain i feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it s top level domain an alternative argument could be made that it should always be grouped under it s parent domain either way it should be consistent
1
67,832
3,282,071,787
IssuesEvent
2015-10-28 02:46:24
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
Clicking the preference button on a filtered deck with "cards selected by relative overdueness" crashes Anki
bug fixedindev Priority-High
Originally reported on Google Code with ID 1967 ``` I create a filtered deck via the desktop client that uses the relative overdueness option. If I click on the preference button inside Ankidroid while opening that deck Ankidroid crashes. ``` Reported by `christian.kleineidam` on 2014-01-15 14:58:46
1.0
Clicking the preference button on a filtered deck with "cards selected by relative overdueness" crashes Anki - Originally reported on Google Code with ID 1967 ``` I create a filtered deck via the desktop client that uses the relative overdueness option. If I click on the preference button inside Ankidroid while opening that deck Ankidroid crashes. ``` Reported by `christian.kleineidam` on 2014-01-15 14:58:46
priority
clicking the preference button on a filtered deck with cards selected by relative overdueness crashes anki originally reported on google code with id i create a filtered deck via the desktop client that uses the relative overdueness option if i click on the preference button inside ankidroid while opening that deck ankidroid crashes reported by christian kleineidam on
1
785,362
27,610,648,899
IssuesEvent
2023-03-09 15:46:19
woocommerce/woocommerce-blocks
https://api.github.com/repos/woocommerce/woocommerce-blocks
opened
Third party tax calculation not displaying after changing shipping method
type: bug priority: high
## Describe the bug When using a third-party tax calculation plugin (AvaTax), they are not displayed on Cart and Checkout blocks after changing shipping methods. A page refresh will properly display them. ## To reproduce Steps to reproduce the behavior: 1. Install Avatax and enable taxes 2. Create several shipping methods 3. Add items to the cart 4. Verify taxes display on the cart and checkout blocks but disappear after changing shipping methods. ## Expected behavior Taxes should be recalculated and displayed after changing shipping methods. Checking out should include taxes. ## Screen recording https://user-images.githubusercontent.com/17236129/224076426-cd37942f-2d3f-47b9-81df-aaade2a3a817.mov ## Additional context Since on-page refresh and on the legacy cart and checkout taxes are correct, we need to investigate if this is a problem that needs fixing on our side.
1.0
Third party tax calculation not displaying after changing shipping method - ## Describe the bug When using a third-party tax calculation plugin (AvaTax), they are not displayed on Cart and Checkout blocks after changing shipping methods. A page refresh will properly display them. ## To reproduce Steps to reproduce the behavior: 1. Install Avatax and enable taxes 2. Create several shipping methods 3. Add items to the cart 4. Verify taxes display on the cart and checkout blocks but disappear after changing shipping methods. ## Expected behavior Taxes should be recalculated and displayed after changing shipping methods. Checking out should include taxes. ## Screen recording https://user-images.githubusercontent.com/17236129/224076426-cd37942f-2d3f-47b9-81df-aaade2a3a817.mov ## Additional context Since on-page refresh and on the legacy cart and checkout taxes are correct, we need to investigate if this is a problem that needs fixing on our side.
priority
third party tax calculation not displaying after changing shipping method describe the bug when using a third party tax calculation plugin avatax they are not displayed on cart and checkout blocks after changing shipping methods a page refresh will properly display them to reproduce steps to reproduce the behavior install avatax and enable taxes create several shipping methods add items to the cart verify taxes display on the cart and checkout blocks but disappear after changing shipping methods expected behavior taxes should be recalculated and displayed after changing shipping methods checking out should include taxes screen recording additional context since on page refresh and on the legacy cart and checkout taxes are correct we need to investigate if this is a problem that needs fixing on our side
1
295,392
9,085,801,102
IssuesEvent
2019-02-18 09:20:50
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
opened
Assertion failure cancelling live data operation using Kafka listener
Component: Live Data Priority: High
### Expected behavior A live data operation should be able to be cancelled cleanly. ### Actual behavior While testing #24799 it was observed that cancelling `MonitorLiveData` fires a debug assertion regarding an empty shared_ptr. Terminal log: ``` ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument ConfigService-[Information] This is Mantid version 3.13.20190218.859 revision g1b57e42 ConfigService-[Information] running on ndw1213 starting 2019-02-18T09:13Z ConfigService-[Information] Properties file(s) loaded: /media/data1/source/github/mantidproject/mantid-builds/nightly-py2/bin/Mantid.properties, /etc/mantid.local.properties, /home/dmn58364/.mantid/Mantid.user.properties ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument FrameworkManager-[Notice] Welcome to Mantid 3.13.20190218.859 FrameworkManager-[Notice] Please cite: http://dx.doi.org/10.1016/j.nima.2014.07.029 and this release: http://dx.doi.org/10.5286/Software/Mantid StartLiveData-[Notice] StartLiveData started LoadLiveData-[Notice] Performing the Replace operation. MonitorLiveData-[Notice] MonitorLiveData started StartLiveData-[Notice] StartLiveData successful, Duration 0.20 seconds python: /usr/include/boost/smart_ptr/shared_ptr.hpp:648: typename boost::detail::sp_member_access<T>::type boost::shared_ptr<T>::operator->() const [with T = Mantid::DataObjects::EventWorkspace; typename boost::detail::sp_member_access<T>::type = Mantid::DataObjects::EventWorkspace*]: Assertion `px != 0' failed. ``` ### Steps to reproduce the behavior * start MantidPlot * set facility to ISIS * load live data * select instrument=MARI * select connection=kafka_event * click run * wait until `MonitorLiveData` has started and click cancel => :boom: ### Platforms affected Verified on Ubuntu 16.04 but I would assume all.
1.0
Assertion failure cancelling live data operation using Kafka listener - ### Expected behavior A live data operation should be able to be cancelled cleanly. ### Actual behavior While testing #24799 it was observed that cancelling `MonitorLiveData` fires a debug assertion regarding an empty shared_ptr. Terminal log: ``` ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument ConfigService-[Information] This is Mantid version 3.13.20190218.859 revision g1b57e42 ConfigService-[Information] running on ndw1213 starting 2019-02-18T09:13Z ConfigService-[Information] Properties file(s) loaded: /media/data1/source/github/mantidproject/mantid-builds/nightly-py2/bin/Mantid.properties, /etc/mantid.local.properties, /home/dmn58364/.mantid/Mantid.user.properties ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument FrameworkManager-[Notice] Welcome to Mantid 3.13.20190218.859 FrameworkManager-[Notice] Please cite: http://dx.doi.org/10.1016/j.nima.2014.07.029 and this release: http://dx.doi.org/10.5286/Software/Mantid StartLiveData-[Notice] StartLiveData started LoadLiveData-[Notice] Performing the Replace operation. MonitorLiveData-[Notice] MonitorLiveData started StartLiveData-[Notice] StartLiveData successful, Duration 0.20 seconds python: /usr/include/boost/smart_ptr/shared_ptr.hpp:648: typename boost::detail::sp_member_access<T>::type boost::shared_ptr<T>::operator->() const [with T = Mantid::DataObjects::EventWorkspace; typename boost::detail::sp_member_access<T>::type = Mantid::DataObjects::EventWorkspace*]: Assertion `px != 0' failed. ``` ### Steps to reproduce the behavior * start MantidPlot * set facility to ISIS * load live data * select instrument=MARI * select connection=kafka_event * click run * wait until `MonitorLiveData` has started and click cancel => :boom: ### Platforms affected Verified on Ubuntu 16.04 but I would assume all.
priority
assertion failure cancelling live data operation using kafka listener expected behavior a live data operation should be able to be cancelled cleanly actual behavior while testing it was observed that cancelling monitorlivedata fires a debug assertion regarding an empty shared ptr terminal log configservice unable to locate directory at etc mantid instrument configservice this is mantid version revision configservice running on starting configservice properties file s loaded media source github mantidproject mantid builds nightly bin mantid properties etc mantid local properties home mantid mantid user properties configservice unable to locate directory at etc mantid instrument frameworkmanager welcome to mantid frameworkmanager please cite and this release startlivedata startlivedata started loadlivedata performing the replace operation monitorlivedata monitorlivedata started startlivedata startlivedata successful duration seconds python usr include boost smart ptr shared ptr hpp typename boost detail sp member access type boost shared ptr operator const assertion px failed steps to reproduce the behavior start mantidplot set facility to isis load live data select instrument mari select connection kafka event click run wait until monitorlivedata has started and click cancel boom platforms affected verified on ubuntu but i would assume all
1
434,477
12,518,897,877
IssuesEvent
2020-06-03 13:40:04
carbon-design-system/ibm-dotcom-library
https://api.github.com/repos/carbon-design-system/ibm-dotcom-library
closed
Create codesandbox for Dotcom Shell (Vanilla)
Airtable Done dev package: vanilla priority: high
<!-- Avoid any type of solutions in this user story --> <!-- replace _{{...}}_ with your own words or remove --> #### User Story <!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} --> > As a `[user role below]`: IBM.com Library adopter > I need to: see the DotcomShell in an isolated codesandbox > so that I can: see how I can easily integrate into my application #### Additional information <!-- {{Please provide any additional information or resources for reference}} --> - Should create under the `ibmdotcom-bot` github user under https://codesandbox.io - 1Password access for `ibmdotcom-bot`: https://ibm.ent.box.com/notes/496014563545 - Template available under codesandbox: https://codesandbox.io/dashboard/templates #### Acceptance criteria - [ ] Codesandbox created for DotcomShell - [ ] Link added to README
1.0
Create codesandbox for Dotcom Shell (Vanilla) - <!-- Avoid any type of solutions in this user story --> <!-- replace _{{...}}_ with your own words or remove --> #### User Story <!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} --> > As a `[user role below]`: IBM.com Library adopter > I need to: see the DotcomShell in an isolated codesandbox > so that I can: see how I can easily integrate into my application #### Additional information <!-- {{Please provide any additional information or resources for reference}} --> - Should create under the `ibmdotcom-bot` github user under https://codesandbox.io - 1Password access for `ibmdotcom-bot`: https://ibm.ent.box.com/notes/496014563545 - Template available under codesandbox: https://codesandbox.io/dashboard/templates #### Acceptance criteria - [ ] Codesandbox created for DotcomShell - [ ] Link added to README
priority
create codesandbox for dotcom shell vanilla user story as a ibm com library adopter i need to see the dotcomshell in an isolated codesandbox so that i can see how i can easily integrate into my application additional information should create under the ibmdotcom bot github user under access for ibmdotcom bot template available under codesandbox acceptance criteria codesandbox created for dotcomshell link added to readme
1
386,304
11,435,062,276
IssuesEvent
2020-02-04 18:38:09
collinbarrett/FilterLists
https://api.github.com/repos/collinbarrett/FilterLists
closed
FilterLists Hub isn't working
bug high priority
I see only dark background on FilterLists Hub and errors on console.
1.0
FilterLists Hub isn't working - I see only dark background on FilterLists Hub and errors on console.
priority
filterlists hub isn t working i see only dark background on filterlists hub and errors on console
1
663,090
22,161,700,081
IssuesEvent
2022-06-04 15:46:11
ApplETS/Notre-Dame
https://api.github.com/repos/ApplETS/Notre-Dame
opened
Rating does not work on button click
bug platform: ios platform: android ready to develop priority: high
**Describe the bug** The button 'rate me' don't do anything. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'More' 2. Click on 'Rate Me' 3. See error **Expected behavior** The button should show the in app review or at least redirect to store listing. **Screenshots** N/A **Additional context** N/A
1.0
Rating does not work on button click - **Describe the bug** The button 'rate me' don't do anything. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'More' 2. Click on 'Rate Me' 3. See error **Expected behavior** The button should show the in app review or at least redirect to store listing. **Screenshots** N/A **Additional context** N/A
priority
rating does not work on button click describe the bug the button rate me don t do anything to reproduce steps to reproduce the behavior go to more click on rate me see error expected behavior the button should show the in app review or at least redirect to store listing screenshots n a additional context n a
1