Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
741,724 | 25,814,585,605 | IssuesEvent | 2022-12-12 03:10:59 | containerd/nerdctl | https://api.github.com/repos/containerd/nerdctl | closed | [Rootful + SELinux] [BuildKit] RUN command throws error code 139 on centos | bug priority/high kind/external | ### Description
Cannot build using containerfile with RUN command on centos. The same containerfile works fine on ubuntu.
### Steps to reproduce the issue
1. Have a simple Containter file:
FROM alpine
RUN ls
2. Build it
### Describe the results you received and expected
nerdctl build -t test:1 .
[+] Building 1.5s (4/5)
[+] Building 1.6s (5/5) FINISHED
=> [internal] load build definition from Containerfile 0.0s
=> => transferring dockerfile: 118B 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 2B 0.0s
=> [internal] load metadata for docker.io/library/alpine:latest 1.0s
=> [1/2] FROM docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.3s
=> => resolve docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.0s
=> => sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 2.81MB / 2.81MB 0.1s
=> => extracting sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 0.1s
=> ERROR [2/2] RUN ls 0.2s
------
> [2/2] RUN ls:
------
Containerfile:2
--------------------
1 | FROM alpine
2 | >>> RUN ls
3 |
--------------------
error: failed to solve: process "/bin/sh -c ls" did not complete successfully: exit code: 139
FATA[0001] unrecognized image format
### What version of nerdctl are you using?
nerdctl version 0.23.0
### Are you using a variant of nerdctl? (e.g., Rancher Desktop)
_No response_
### Host information
Client:
Namespace: default
Debug Mode: false
Server:
Server Version: v1.6.8
Storage Driver: overlayfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Cgroup Version: 1
Plugins:
Log: fluentd journald json-file
Storage: native overlayfs
Security Options:
seccomp
Profile: default
Kernel Version: 4.18.0-408.el8.x86_64
Operating System: CentOS Stream 8
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 14.47GiB
Name: pcr-nerdctl-tst2
ID: 2908f464-ed9f-4476-99cc-205a294e6e51
WARNING: bridge-nf-call-iptables is disabled
WARNING: bridge-nf-call-ip6tables is disabled
| 1.0 | [Rootful + SELinux] [BuildKit] RUN command throws error code 139 on centos - ### Description
Cannot build using containerfile with RUN command on centos. The same containerfile works fine on ubuntu.
### Steps to reproduce the issue
1. Have a simple Containter file:
FROM alpine
RUN ls
2. Build it
### Describe the results you received and expected
nerdctl build -t test:1 .
[+] Building 1.5s (4/5)
[+] Building 1.6s (5/5) FINISHED
=> [internal] load build definition from Containerfile 0.0s
=> => transferring dockerfile: 118B 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 2B 0.0s
=> [internal] load metadata for docker.io/library/alpine:latest 1.0s
=> [1/2] FROM docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.3s
=> => resolve docker.io/library/alpine@sha256:bc41182d7ef5ffc53a40b044e725193bc10142a1243f395ee852a8d9730fc2ad 0.0s
=> => sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 2.81MB / 2.81MB 0.1s
=> => extracting sha256:213ec9aee27d8be045c6a92b7eac22c9a64b44558193775a1a7f626352392b49 0.1s
=> ERROR [2/2] RUN ls 0.2s
------
> [2/2] RUN ls:
------
Containerfile:2
--------------------
1 | FROM alpine
2 | >>> RUN ls
3 |
--------------------
error: failed to solve: process "/bin/sh -c ls" did not complete successfully: exit code: 139
FATA[0001] unrecognized image format
### What version of nerdctl are you using?
nerdctl version 0.23.0
### Are you using a variant of nerdctl? (e.g., Rancher Desktop)
_No response_
### Host information
Client:
Namespace: default
Debug Mode: false
Server:
Server Version: v1.6.8
Storage Driver: overlayfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Cgroup Version: 1
Plugins:
Log: fluentd journald json-file
Storage: native overlayfs
Security Options:
seccomp
Profile: default
Kernel Version: 4.18.0-408.el8.x86_64
Operating System: CentOS Stream 8
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 14.47GiB
Name: pcr-nerdctl-tst2
ID: 2908f464-ed9f-4476-99cc-205a294e6e51
WARNING: bridge-nf-call-iptables is disabled
WARNING: bridge-nf-call-ip6tables is disabled
| priority | run command throws error code on centos description cannot build using containerfile with run command on centos the same containerfile works fine on ubuntu steps to reproduce the issue have a simple containter file from alpine run ls build it describe the results you received and expected nerdctl build t test building building finished load build definition from containerfile transferring dockerfile load dockerignore transferring context load metadata for docker io library alpine latest from docker io library alpine resolve docker io library alpine extracting error run ls run ls containerfile from alpine run ls error failed to solve process bin sh c ls did not complete successfully exit code fata unrecognized image format what version of nerdctl are you using nerdctl version are you using a variant of nerdctl e g rancher desktop no response host information client namespace default debug mode false server server version storage driver overlayfs logging driver json file cgroup driver cgroupfs cgroup version plugins log fluentd journald json file storage native overlayfs security options seccomp profile default kernel version operating system centos stream ostype linux architecture cpus total memory name pcr nerdctl id warning bridge nf call iptables is disabled warning bridge nf call is disabled | 1 |
566,104 | 16,796,077,732 | IssuesEvent | 2021-06-16 03:52:40 | sodafoundation/multi-cloud | https://api.github.com/repos/sodafoundation/multi-cloud | opened | AccessKey , SecretKey to be created from API | High Priority | **Issue/Feature Description:**
AccessKey , SecretKey to be created from API - To enable users to create AKSK from the API , rather than the UI.
AkSk from API will enable users to create the AKSK from the API and integrate to their Clients.
**Why this issue to fixed / feature is needed(give scenarios or use cases):**
AkSk from API will enable users to create the AKSK from the API and integrate to their Clients.
**How to reproduce, in case of a bug:**
N/A
**Other Notes / Environment Information: (Please give the env information, log link or any useful information for this issue)**
| 1.0 | AccessKey , SecretKey to be created from API - **Issue/Feature Description:**
AccessKey , SecretKey to be created from API - To enable users to create AKSK from the API , rather than the UI.
AkSk from API will enable users to create the AKSK from the API and integrate to their Clients.
**Why this issue to fixed / feature is needed(give scenarios or use cases):**
AkSk from API will enable users to create the AKSK from the API and integrate to their Clients.
**How to reproduce, in case of a bug:**
N/A
**Other Notes / Environment Information: (Please give the env information, log link or any useful information for this issue)**
| priority | accesskey secretkey to be created from api issue feature description accesskey secretkey to be created from api to enable users to create aksk from the api rather than the ui aksk from api will enable users to create the aksk from the api and integrate to their clients why this issue to fixed feature is needed give scenarios or use cases aksk from api will enable users to create the aksk from the api and integrate to their clients how to reproduce in case of a bug n a other notes environment information please give the env information log link or any useful information for this issue | 1 |
96,091 | 3,964,411,317 | IssuesEvent | 2016-05-03 00:45:32 | daronco/test-issue-migrate2 | https://api.github.com/repos/daronco/test-issue-migrate2 | closed | Set a numeric voiceBridge for every room | Priority: High Status: Resolved Type: Bug | ---
Author Name: **Leonardo Daronco** (@daronco)
Original Redmine Issue: 145, http://dev.mconf.org/redmine/issues/145
Original Assignee: Leonardo Daronco
---
@Google user: leonardo...@gmail.com@
Use a numeric "voiceBridge" param when creating a room.
From the BigBlueButton API documentation: "we recommend you always pass a 5 digit voiceBridge parameter -- and have it begin with the digit '7' if you are using the default FreeSWITCH setup"
So the voiceBridge should have 5 digits, the first one being a '7' and the last ones can be taken from the meetingID (after implementing "/p/mconf/issues/detail?id=38": issue #38 ) or can be random (but unique).
| 1.0 | Set a numeric voiceBridge for every room - ---
Author Name: **Leonardo Daronco** (@daronco)
Original Redmine Issue: 145, http://dev.mconf.org/redmine/issues/145
Original Assignee: Leonardo Daronco
---
@Google user: leonardo...@gmail.com@
Use a numeric "voiceBridge" param when creating a room.
From the BigBlueButton API documentation: "we recommend you always pass a 5 digit voiceBridge parameter -- and have it begin with the digit '7' if you are using the default FreeSWITCH setup"
So the voiceBridge should have 5 digits, the first one being a '7' and the last ones can be taken from the meetingID (after implementing "/p/mconf/issues/detail?id=38": issue #38 ) or can be random (but unique).
| priority | set a numeric voicebridge for every room author name leonardo daronco daronco original redmine issue original assignee leonardo daronco google user leonardo gmail com use a numeric voicebridge param when creating a room from the bigbluebutton api documentation we recommend you always pass a digit voicebridge parameter and have it begin with the digit if you are using the default freeswitch setup so the voicebridge should have digits the first one being a and the last ones can be taken from the meetingid after implementing p mconf issues detail id issue or can be random but unique | 1 |
425,104 | 12,335,783,449 | IssuesEvent | 2020-05-14 12:34:46 | zeebe-io/zeebe | https://api.github.com/repos/zeebe-io/zeebe | opened | Possible regression in job activation | Impact: Performance Scope: broker Severity: High Status: Needs Priority Type: Bug | **Describe the bug**
I observed a big drop in throughput on running our normal benchmark.

I would normally expect ~200 workflows and task to be completed.
It seems that job activation is the problem since the activation latency is quite high.

The standalone gateway also throws endless the following timeouts:
```
2020-05-14 12:31:23.930 [io.zeebe.gateway.impl.broker.BrokerRequestManager] [gateway-scheduler-zb-actors-0] ERROR io.zeebe.gateway - Error handling gRPC request
io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: Time out between gateway and broker: Request type command-api-1 timed out in 15000 milliseconds
at io.grpc.Status.asRuntimeException(Status.java:524) ~[grpc-api-1.29.0.jar:1.29.0]
at io.zeebe.gateway.EndpointManager.convertThrowable(EndpointManager.java:397) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.EndpointManager.lambda$sendRequest$3(EndpointManager.java:311) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequest$3(BrokerRequestManager.java:148) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequestInternal$5(BrokerRequestManager.java:191) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.future.FutureContinuationRunnable.run(FutureContinuationRunnable.java:33) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorJob.invoke(ActorJob.java:76) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorJob.execute(ActorJob.java:39) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorTask.execute(ActorTask.java:118) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:107) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.doWork(ActorThread.java:91) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.run(ActorThread.java:204) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
Caused by: java.util.concurrent.TimeoutException: Request type command-api-1 timed out in 15000 milliseconds
at io.atomix.cluster.messaging.impl.AbstractClientConnection$Callback.timeout(AbstractClientConnection.java:163) ~[atomix-cluster-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) ~[?:?]
at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown Source) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) ~[?:?]
at java.lang.Thread.run(Unknown Source) ~[?:?]
```
**To Reproduce**
Run the helm chart (v100) with our benchmark,
**Expected behavior**
Around 200 workflows are completed per second.
| 1.0 | Possible regression in job activation - **Describe the bug**
I observed a big drop in throughput on running our normal benchmark.

I would normally expect ~200 workflows and task to be completed.
It seems that job activation is the problem since the activation latency is quite high.

The standalone gateway also throws endless the following timeouts:
```
2020-05-14 12:31:23.930 [io.zeebe.gateway.impl.broker.BrokerRequestManager] [gateway-scheduler-zb-actors-0] ERROR io.zeebe.gateway - Error handling gRPC request
io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: Time out between gateway and broker: Request type command-api-1 timed out in 15000 milliseconds
at io.grpc.Status.asRuntimeException(Status.java:524) ~[grpc-api-1.29.0.jar:1.29.0]
at io.zeebe.gateway.EndpointManager.convertThrowable(EndpointManager.java:397) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.EndpointManager.lambda$sendRequest$3(EndpointManager.java:311) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequest$3(BrokerRequestManager.java:148) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.gateway.impl.broker.BrokerRequestManager.lambda$sendRequestInternal$5(BrokerRequestManager.java:191) ~[zeebe-gateway-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.future.FutureContinuationRunnable.run(FutureContinuationRunnable.java:33) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorJob.invoke(ActorJob.java:76) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorJob.execute(ActorJob.java:39) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorTask.execute(ActorTask.java:118) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:107) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.doWork(ActorThread.java:91) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at io.zeebe.util.sched.ActorThread.run(ActorThread.java:204) [zeebe-util-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
Caused by: java.util.concurrent.TimeoutException: Request type command-api-1 timed out in 15000 milliseconds
at io.atomix.cluster.messaging.impl.AbstractClientConnection$Callback.timeout(AbstractClientConnection.java:163) ~[atomix-cluster-0.24.0-SNAPSHOT.jar:0.24.0-SNAPSHOT]
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) ~[?:?]
at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown Source) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) ~[?:?]
at java.lang.Thread.run(Unknown Source) ~[?:?]
```
**To Reproduce**
Run the helm chart (v100) with our benchmark,
**Expected behavior**
Around 200 workflows are completed per second.
| priority | possible regression in job activation describe the bug i observed a big drop in throughput on running our normal benchmark i would normally expect workflows and task to be completed it seems that job activation is the problem since the activation latency is quite high the standalone gateway also throws endless the following timeouts error io zeebe gateway error handling grpc request io grpc statusruntimeexception deadline exceeded time out between gateway and broker request type command api timed out in milliseconds at io grpc status asruntimeexception status java at io zeebe gateway endpointmanager convertthrowable endpointmanager java at io zeebe gateway endpointmanager lambda sendrequest endpointmanager java at io zeebe gateway impl broker brokerrequestmanager lambda sendrequest brokerrequestmanager java at io zeebe gateway impl broker brokerrequestmanager lambda sendrequestinternal brokerrequestmanager java at io zeebe util sched future futurecontinuationrunnable run futurecontinuationrunnable java at io zeebe util sched actorjob invoke actorjob java at io zeebe util sched actorjob execute actorjob java at io zeebe util sched actortask execute actortask java at io zeebe util sched actorthread executecurrenttask actorthread java at io zeebe util sched actorthread dowork actorthread java at io zeebe util sched actorthread run actorthread java caused by java util concurrent timeoutexception request type command api timed out in milliseconds at io atomix cluster messaging impl abstractclientconnection callback timeout abstractclientconnection java at java util concurrent executors runnableadapter call unknown source at java util concurrent futuretask run unknown source at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run unknown source at java util concurrent threadpoolexecutor runworker unknown source at java util concurrent threadpoolexecutor worker run unknown source at java lang thread run unknown source to reproduce run the helm chart with our benchmark expected behavior around workflows are completed per second | 1 |
619,804 | 19,535,416,577 | IssuesEvent | 2021-12-31 05:08:00 | ajayyy/SponsorBlock | https://api.github.com/repos/ajayyy/SponsorBlock | closed | Better-sqlite3 and other improvements | HIGH PRIORITY | [better-sqlite](https://github.com/JoshuaWise/better-sqlite3) is faster than the more popular `sqlite3` node library especially when load increases. The reason is that sqlite3 is "asynchronous" even though all the computation still happens on the same thread. This means the only thing it being async does is it increases the overhead significantly. (The reason it's more popular is as far as I know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes.) | 1.0 | Better-sqlite3 and other improvements - [better-sqlite](https://github.com/JoshuaWise/better-sqlite3) is faster than the more popular `sqlite3` node library especially when load increases. The reason is that sqlite3 is "asynchronous" even though all the computation still happens on the same thread. This means the only thing it being async does is it increases the overhead significantly. (The reason it's more popular is as far as I know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes.) | priority | better and other improvements is faster than the more popular node library especially when load increases the reason is that is asynchronous even though all the computation still happens on the same thread this means the only thing it being async does is it increases the overhead significantly the reason it s more popular is as far as i know only because node devs fear stuff that is synchronous without really understanding what happens behind the scenes | 1 |
523,524 | 15,184,319,187 | IssuesEvent | 2021-02-15 09:24:54 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | [Formatter] Formatting doesn't work properly when a part of the string is in a different line | Area/Formatting Priority/High Team/Tooling Type/Bug | **Description:**
Please consider the following scenario

When I format the above code snippet additional spaces are added to the string values as given below.

**Steps to reproduce:**
**Affected Versions:**
SLP9-snapshot
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | [Formatter] Formatting doesn't work properly when a part of the string is in a different line - **Description:**
Please consider the following scenario

When I format the above code snippet additional spaces are added to the string values as given below.

**Steps to reproduce:**
**Affected Versions:**
SLP9-snapshot
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | formatting doesn t work properly when a part of the string is in a different line description please consider the following scenario when i format the above code snippet additional spaces are added to the string values as given below steps to reproduce affected versions snapshot os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 1 |
393,662 | 11,623,112,634 | IssuesEvent | 2020-02-27 08:15:58 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Map water doesnt not always load on web interface | Priority: High Status: Fixed | Repeatedly reload any page with a map, sometimes the water does not load correctly.
https://gyazo.com/ce4aae1286389ebf4692477e4c662d25 | 1.0 | Map water doesnt not always load on web interface - Repeatedly reload any page with a map, sometimes the water does not load correctly.
https://gyazo.com/ce4aae1286389ebf4692477e4c662d25 | priority | map water doesnt not always load on web interface repeatedly reload any page with a map sometimes the water does not load correctly | 1 |
65,004 | 3,222,593,846 | IssuesEvent | 2015-10-09 02:33:50 | cs2103aug2015-t11-4j/main | https://api.github.com/repos/cs2103aug2015-t11-4j/main | closed | Parser support for update task | :parser priority.high type.task | Takes in string value from Logic and edit the certain task object and return the object to Logic | 1.0 | Parser support for update task - Takes in string value from Logic and edit the certain task object and return the object to Logic | priority | parser support for update task takes in string value from logic and edit the certain task object and return the object to logic | 1 |
398,475 | 11,741,500,717 | IssuesEvent | 2020-03-11 21:54:13 | SacredDuckwhale/Rarity | https://api.github.com/repos/SacredDuckwhale/Rarity | opened | The Combat-log based attempts detection doesn't work properly when used to detect outdoor world bosses | complexity: moderate module:core priority:high status:accepted type:bug | Verified for Dunegorger Kraulok. After looking at the code, it appears any UNIT_DIED event is now triggering the detection, when only those that are caused by the player or their party/raid should count.
The implementation appears to rely on bit flags for a certain srcFlag bitmap field (see https://wow.gamepedia.com/API_CombatLogGetCurrentEventInfo), but those are no longer working? The value is always -2147483648 and so ALL kills count regardless of who actually caused them.
Potential solutions:
* Repair the bit flag detection, if Blizzard hasn't removed/broken it
* Failing that, it might be possible to rely on the defeat detection (loot lockout) and count an attempt if and only if the player isn't yet logged out
Notes:
* MOP world bosses are probably affected, as well | 1.0 | The Combat-log based attempts detection doesn't work properly when used to detect outdoor world bosses - Verified for Dunegorger Kraulok. After looking at the code, it appears any UNIT_DIED event is now triggering the detection, when only those that are caused by the player or their party/raid should count.
The implementation appears to rely on bit flags for a certain srcFlag bitmap field (see https://wow.gamepedia.com/API_CombatLogGetCurrentEventInfo), but those are no longer working? The value is always -2147483648 and so ALL kills count regardless of who actually caused them.
Potential solutions:
* Repair the bit flag detection, if Blizzard hasn't removed/broken it
* Failing that, it might be possible to rely on the defeat detection (loot lockout) and count an attempt if and only if the player isn't yet logged out
Notes:
* MOP world bosses are probably affected, as well | priority | the combat log based attempts detection doesn t work properly when used to detect outdoor world bosses verified for dunegorger kraulok after looking at the code it appears any unit died event is now triggering the detection when only those that are caused by the player or their party raid should count the implementation appears to rely on bit flags for a certain srcflag bitmap field see but those are no longer working the value is always and so all kills count regardless of who actually caused them potential solutions repair the bit flag detection if blizzard hasn t removed broken it failing that it might be possible to rely on the defeat detection loot lockout and count an attempt if and only if the player isn t yet logged out notes mop world bosses are probably affected as well | 1 |
810,374 | 30,239,070,418 | IssuesEvent | 2023-07-06 12:22:40 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Cypress image snapshots not working | Bug :lady_beetle: Sprint Priority: High Frontend :sunglasses: | **Describe the bug**
Cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the UI.
**To Reproduce**
Steps to reproduce the behavior:
- In the new translations UI, in the component for the translations lists, change the type for the action buttons, so that they visually change.
- Run the E2E relevant to translations.
- There’s no error for the visual change.
**Expected behavior**
We should have a very low threshold of tolerance for visual changes in the UI and the test should fail.
| 1.0 | Cypress image snapshots not working - **Describe the bug**
Cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the UI.
**To Reproduce**
Steps to reproduce the behavior:
- In the new translations UI, in the component for the translations lists, change the type for the action buttons, so that they visually change.
- Run the E2E relevant to translations.
- There’s no error for the visual change.
**Expected behavior**
We should have a very low threshold of tolerance for visual changes in the UI and the test should fail.
| priority | cypress image snapshots not working describe the bug cypress image snapshots are likely to be misconfigured since they are not properly reporting changes to the ui to reproduce steps to reproduce the behavior in the new translations ui in the component for the translations lists change the type for the action buttons so that they visually change run the relevant to translations there’s no error for the visual change expected behavior we should have a very low threshold of tolerance for visual changes in the ui and the test should fail | 1 |
423,632 | 12,299,364,335 | IssuesEvent | 2020-05-11 12:15:21 | bounswe/bounswe2020group1 | https://api.github.com/repos/bounswe/bounswe2020group1 | closed | Implement the functions for "find similar words" API | priority:high type:implementation | Implementing functions that use the API provided by Datamuse. | 1.0 | Implement the functions for "find similar words" API - Implementing functions that use the API provided by Datamuse. | priority | implement the functions for find similar words api implementing functions that use the api provided by datamuse | 1 |
808,195 | 30,037,669,465 | IssuesEvent | 2023-06-27 13:43:39 | tum-esm/hermes | https://api.github.com/repos/tum-esm/hermes | closed | Send in-flow sensor measurement with every CO2 measurement over MQTT | type:feature status:implemented high-priority scope:sensor | The frequency of the in-flow measurements needs to be higher. (Current implementation every 2 minutes)
Change to directly integrate the in-flow sensor data into the CO2 measurement MQTT data stream. This also makes it more visible what sensor data the CO2 sensor received to perform the correction and allows us to do it ourselves in the future.
Physical Background: The pump creates a flow between 0.4-0.6 ppm. Depending on the flow the underpressure in the system changes. This change has a direct influence on the correction based on pressure in the Vaisala GMP343. | 1.0 | Send in-flow sensor measurement with every CO2 measurement over MQTT - The frequency of the in-flow measurements needs to be higher. (Current implementation every 2 minutes)
Change to directly integrate the in-flow sensor data into the CO2 measurement MQTT data stream. This also makes it more visible what sensor data the CO2 sensor received to perform the correction and allows us to do it ourselves in the future.
Physical Background: The pump creates a flow between 0.4-0.6 ppm. Depending on the flow the underpressure in the system changes. This change has a direct influence on the correction based on pressure in the Vaisala GMP343. | priority | send in flow sensor measurement with every measurement over mqtt the frequency of the in flow measurements needs to be higher current implementation every minutes change to directly integrate the in flow sensor data into the measurement mqtt data stream this also makes it more visible what sensor data the sensor received to perform the correction and allows us to do it ourselves in the future physical background the pump creates a flow between ppm depending on the flow the underpressure in the system changes this change has a direct influence on the correction based on pressure in the vaisala | 1 |
792,351 | 27,956,882,081 | IssuesEvent | 2023-03-24 13:03:33 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Needs to get the default value of an input parameter in a resource function | Type/Improvement Priority/High Team/jBallerina | **Description:**
When there is an input parameter with a default value in a resource function, there should be a way to get the default value of this parameter. This is a requirement for the Ballerina GraphQL package implementation.
Related: #27417, https://github.com/ballerina-platform/ballerina-standard-library/issues/1266 | 1.0 | Needs to get the default value of an input parameter in a resource function - **Description:**
When there is an input parameter with a default value in a resource function, there should be a way to get the default value of this parameter. This is a requirement for the Ballerina GraphQL package implementation.
Related: #27417, https://github.com/ballerina-platform/ballerina-standard-library/issues/1266 | priority | needs to get the default value of an input parameter in a resource function description when there is an input parameter with a default value in a resource function there should be a way to get the default value of this parameter this is a requirement for the ballerina graphql package implementation related | 1 |
484,359 | 13,938,471,195 | IssuesEvent | 2020-10-22 15:18:24 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Pattern based anomaly detection engine for the cloud | Complexity/High Component/Analytics Priority/Low gateway research | **Is your feature request related to a problem? Please describe.**
$subject allows monitoring an Identity Server cloud deployment for unusual behaviors, based on pattern-based anomalies of various signals coming from the deployment. The engine can be used to trigger alerts in near-realtime, on potential malicious requests, attacks, and other unexpected activities.
**Describe the solution you would prefer**
Come up with an anomaly detection engine PoC, where various signals from the Identity Server deployment(request rate, authentication events, etc) can be analyzed in general, and compare it with the patterns observed in the past, to identify anomalies. The engine should support various un-supervised analysis mechanisms, due to the vast amount of data that can be generated from a cloud deployment.
**Additional context**
-
| 1.0 | Pattern based anomaly detection engine for the cloud - **Is your feature request related to a problem? Please describe.**
$subject allows monitoring an Identity Server cloud deployment for unusual behaviors, based on pattern-based anomalies of various signals coming from the deployment. The engine can be used to trigger alerts in near-realtime, on potential malicious requests, attacks, and other unexpected activities.
**Describe the solution you would prefer**
Come up with an anomaly detection engine PoC, where various signals from the Identity Server deployment(request rate, authentication events, etc) can be analyzed in general, and compare it with the patterns observed in the past, to identify anomalies. The engine should support various un-supervised analysis mechanisms, due to the vast amount of data that can be generated from a cloud deployment.
**Additional context**
-
| priority | pattern based anomaly detection engine for the cloud is your feature request related to a problem please describe subject allows monitoring an identity server cloud deployment for unusual behaviors based on pattern based anomalies of various signals coming from the deployment the engine can be used to trigger alerts in near realtime on potential malicious requests attacks and other unexpected activities describe the solution you would prefer come up with an anomaly detection engine poc where various signals from the identity server deployment request rate authentication events etc can be analyzed in general and compare it with the patterns observed in the past to identify anomalies the engine should support various un supervised analysis mechanisms due to the vast amount of data that can be generated from a cloud deployment additional context | 1 |
184,914 | 6,717,386,738 | IssuesEvent | 2017-10-14 20:35:20 | semperfiwebdesign/all-in-one-seo-pack | https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack | opened | Exclude Pages option no longer working | Initial Review Priority | High | As reported by Albert Belzer (belzer9@gmail.com) on October 14, 2017.
In his original e-mail, Albert states that both the "Disable SEO for this post/page option" setting on the Edit screen and the "Exlude Pages" setting in the General Settings are not working.
I checked this and although the former is working, the latter seems like it's not. | 1.0 | Exclude Pages option no longer working - As reported by Albert Belzer (belzer9@gmail.com) on October 14, 2017.
In his original e-mail, Albert states that both the "Disable SEO for this post/page option" setting on the Edit screen and the "Exlude Pages" setting in the General Settings are not working.
I checked this and although the former is working, the latter seems like it's not. | priority | exclude pages option no longer working as reported by albert belzer gmail com on october in his original e mail albert states that both the disable seo for this post page option setting on the edit screen and the exlude pages setting in the general settings are not working i checked this and although the former is working the latter seems like it s not | 1 |
382,199 | 11,302,244,285 | IssuesEvent | 2020-01-17 17:12:56 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | reopened | Prototype APIs can't be used for API products | 3.1.0 Priority/Highest Resolution/Fixed Type/Bug | $subject.
The inline script is not copied to the API product synapse. Hense invocation fails.
| 1.0 | Prototype APIs can't be used for API products - $subject.
The inline script is not copied to the API product synapse. Hense invocation fails.
| priority | prototype apis can t be used for api products subject the inline script is not copied to the api product synapse hense invocation fails | 1 |
382,645 | 11,309,760,056 | IssuesEvent | 2020-01-19 15:16:11 | FederatedAI/FATE | https://api.github.com/repos/FederatedAI/FATE | closed | Support for secret sharing scheme | enhancement priority:high research | **Is your feature request related to a problem? Please describe.**
Secret sharing Scheme is a must-have for FATE project.
**Describe the solution you'd like**
Do R&D on implementing secret sharing operations, such as:
1. Create beaver triple
2. Add, Multiply, Division, Compare, and others
Having secret sharing operations been created, then
1. Implement secret sharing based LR
2. Implement secret sharing based FTL
These works do not need to be full-fledged for industrial applications. However, they should be able to help us create various secure federated learning algorithms/prototypes.
| 1.0 | Support for secret sharing scheme - **Is your feature request related to a problem? Please describe.**
Secret sharing Scheme is a must-have for FATE project.
**Describe the solution you'd like**
Do R&D on implementing secret sharing operations, such as:
1. Create beaver triple
2. Add, Multiply, Division, Compare, and others
Having secret sharing operations been created, then
1. Implement secret sharing based LR
2. Implement secret sharing based FTL
These works do not need to be full-fledged for industrial applications. However, they should be able to help us create various secure federated learning algorithms/prototypes.
| priority | support for secret sharing scheme is your feature request related to a problem please describe secret sharing scheme is a must have for fate project describe the solution you d like do r d on implementing secret sharing operations such as create beaver triple add multiply division compare and others having secret sharing operations been created then implement secret sharing based lr implement secret sharing based ftl these works do not need to be full fledged for industrial applications however they should be able to help us create various secure federated learning algorithms prototypes | 1 |
72,470 | 3,386,257,735 | IssuesEvent | 2015-11-27 16:22:22 | CosmosOS/Cosmos | https://api.github.com/repos/CosmosOS/Cosmos | closed | Dup tries to pop more stuff from analytical stack than there is! | area_compiler complexity_medium pending_verification priority_high | Log:
```
4> Error: Exception: System.Exception: Error compiling method 'SystemVoidKernelCommandsInputCommand': System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is!
4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369
4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714
4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 ---> System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is!
4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369
4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714
4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514
4> --- End of inner exception stack trace ---
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 529
4> at Cosmos.IL2CPU.ILScanner.Assemble() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 944
4> at Cosmos.IL2CPU.ILScanner.Execute(MethodBase aStartMethod) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 256
4> at Cosmos.IL2CPU.CompilerEngine.Execute() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\CompilerEngine.cs:line 238
```
And there is code where it heappen:
```C#
public static void InputCommand()
{
Console.Write("D:/command>");
comd = Console.ReadLine();
comd = comd.ToLower();
if (comd == "reboot") h.Power.Restart();
else if (comd == "shutdown") h.Power.Shutdown();
else if (comd == "echo")
{
Console.Write("Echo>");
arg = Console.ReadLine();
Console.WriteLine(arg);
}
else if (comd == "notepad")
{
System.CLI.Applications.Notepad();
}
else if (comd == "cls")
{
Console.Clear();
Console.WriteLine("TriangleOS");
Console.WriteLine("=============================");
}
else if (comd == "soundtest")
{
Console.Write("Frequency>");
arg = Console.ReadLine();
Console.Write("Duration>");
optarg = Console.ReadLine();
Console.Write("Eh, this isn't implemted right now.");
//h.Multimedia.Speakers.CallSound(int.Parse(arg), int.Parse(optarg));
}
else if (comd == "boot")
{
Console.WriteLine("Starting TriangleOS.Drivers . . .");
//ProcessManager.Process Audio = new ProcessManager.Process();
//ProcessManager.Process Graphics = new ProcessManager.Process();
//Graphics.ProcessThread = new System.Threading.Thread(
h.Graphics.LowLevel.init();
//);
//ProcessManager.Process Mouse = new ProcessManager.Process();
//Mouse.ProcessThread = new System.Threading.Thread(
h.Mouse.InitMouse();
//);
//Graphics.Start();
//Mouse.Start();
//Audio.ProcessThread = new System.Threading.Thread(
h.Multimedia.Speakers.IntailizeAudio();
//);
Kernel.GUI();
}
else if (comd == "cliboot")
{
System.CLI.Controls.TextBox Text = new System.CLI.Controls.TextBox();
Text.x = 1;
Text.y = 1;
Text.length = 20;
Text.DrawTextBox();
Text.TypeInto();
System.CLI.Controls.Button OK = new System.CLI.Controls.Button();
OK.y = 22;
OK.x = 1;
OK.width = 6;
OK.height = 1;
OK.text = "OK";
OK.DrawButton();
}
else if (comd == "calculator")
{
System.CLI.Applications.Calculator();
}
else if (comd == "cd")
{
h.Graphics.Console.ErrO("Impossible operation performed. Can't request I/O while it isn't running!");
}
else if (comd == "dir")
{
Console.WriteLine("This isn't folder. you can use <cd> to go up folder.");
}
else if (comd == "paint")
{
System.CLI.Applications.Paint();
}
else if (comd == "changelog")
{
Console.WriteLine("You are running version 0.0.3 Dev. Only Luka see the DEV!");
Console.WriteLine("v0.0.2:");
Console.WriteLine("Blue screen with cursor. not clearing.");
Console.WriteLine("v0.0.3:");
Console.WriteLine("Command Line Shell with Broken CLI, but working unresponsive GUI, but with DOS-like Shell. I/O, Audio, Multithreading, Shutdown doesn't work.");
}
else if (comd == "help")
{
Console.WriteLine("Copyright 2015 Thontelix TriangleOS. Special thanks to Cosmos .net ASM Compiler.");
Console.WriteLine("type CHANGELOG to get version changes.");
Console.WriteLine("How to use:");
Console.WriteLine("After every typed command, press enter. the output of command will be detailed. If there is blinking bottom line cursor, then you need to input something, if it doesn't give any feedback, then its operating a activity. If you want GUI, type 'boot' and press enter.");
Console.WriteLine("Commands:");
Console.WriteLine("shutdown - Gives you ability to safe turn off PC");
Console.WriteLine("boot - Boots you into TriangleOS");
Console.WriteLine("reboot - Reboots your PC");
Console.WriteLine("cd - Travel through directories");
Console.WriteLine("dir - Read content of directory");
Console.WriteLine("echo - Backs string you enter");
Console.WriteLine("shutdown - Backs string you enter");
Console.WriteLine("soundtest - Speakers Drivers. Doesnt work for now");
Console.WriteLine("help - Gives you list of commands.");
Console.WriteLine("paint - Console Paint App. CAUTION:Not for epilepsy persons.");
}
else
{
Console.WriteLine("That command doesn't exist. :(");
}
}
``` | 1.0 | Dup tries to pop more stuff from analytical stack than there is! - Log:
```
4> Error: Exception: System.Exception: Error compiling method 'SystemVoidKernelCommandsInputCommand': System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is!
4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369
4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714
4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514 ---> System.Exception: OpCode IL_014D: Dup tries to pop more stuff from analytical stack than there is!
4> at Cosmos.IL2CPU.ILOpCode.InterpretStackTypes(IDictionary`2 aOpCodes, Stack`1 aStack, Boolean& aSituationChanged, Int32 aMaxRecursionDepth) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILOpCode.cs:line 369
4> at Cosmos.IL2CPU.AppAssembler.InterpretInstructionsToDetermineStackTypes(List`1 aCurrentGroup) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 714
4> at Cosmos.IL2CPU.AppAssembler.EmitInstructions(MethodInfo aMethod, List`1 aCurrentGroup, Boolean& emitINT3) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 557
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 514
4> --- End of inner exception stack trace ---
4> at Cosmos.IL2CPU.AppAssembler.ProcessMethod(MethodInfo aMethod, List`1 aOpCodes) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\AppAssembler.cs:line 529
4> at Cosmos.IL2CPU.ILScanner.Assemble() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 944
4> at Cosmos.IL2CPU.ILScanner.Execute(MethodBase aStartMethod) in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\ILScanner.cs:line 256
4> at Cosmos.IL2CPU.CompilerEngine.Execute() in C:\Users\Luka\Desktop\Cosmos-master\source\Cosmos.IL2CPU\CompilerEngine.cs:line 238
```
And there is code where it heappen:
```C#
public static void InputCommand()
{
Console.Write("D:/command>");
comd = Console.ReadLine();
comd = comd.ToLower();
if (comd == "reboot") h.Power.Restart();
else if (comd == "shutdown") h.Power.Shutdown();
else if (comd == "echo")
{
Console.Write("Echo>");
arg = Console.ReadLine();
Console.WriteLine(arg);
}
else if (comd == "notepad")
{
System.CLI.Applications.Notepad();
}
else if (comd == "cls")
{
Console.Clear();
Console.WriteLine("TriangleOS");
Console.WriteLine("=============================");
}
else if (comd == "soundtest")
{
Console.Write("Frequency>");
arg = Console.ReadLine();
Console.Write("Duration>");
optarg = Console.ReadLine();
Console.Write("Eh, this isn't implemted right now.");
//h.Multimedia.Speakers.CallSound(int.Parse(arg), int.Parse(optarg));
}
else if (comd == "boot")
{
Console.WriteLine("Starting TriangleOS.Drivers . . .");
//ProcessManager.Process Audio = new ProcessManager.Process();
//ProcessManager.Process Graphics = new ProcessManager.Process();
//Graphics.ProcessThread = new System.Threading.Thread(
h.Graphics.LowLevel.init();
//);
//ProcessManager.Process Mouse = new ProcessManager.Process();
//Mouse.ProcessThread = new System.Threading.Thread(
h.Mouse.InitMouse();
//);
//Graphics.Start();
//Mouse.Start();
//Audio.ProcessThread = new System.Threading.Thread(
h.Multimedia.Speakers.IntailizeAudio();
//);
Kernel.GUI();
}
else if (comd == "cliboot")
{
System.CLI.Controls.TextBox Text = new System.CLI.Controls.TextBox();
Text.x = 1;
Text.y = 1;
Text.length = 20;
Text.DrawTextBox();
Text.TypeInto();
System.CLI.Controls.Button OK = new System.CLI.Controls.Button();
OK.y = 22;
OK.x = 1;
OK.width = 6;
OK.height = 1;
OK.text = "OK";
OK.DrawButton();
}
else if (comd == "calculator")
{
System.CLI.Applications.Calculator();
}
else if (comd == "cd")
{
h.Graphics.Console.ErrO("Impossible operation performed. Can't request I/O while it isn't running!");
}
else if (comd == "dir")
{
Console.WriteLine("This isn't folder. you can use <cd> to go up folder.");
}
else if (comd == "paint")
{
System.CLI.Applications.Paint();
}
else if (comd == "changelog")
{
Console.WriteLine("You are running version 0.0.3 Dev. Only Luka see the DEV!");
Console.WriteLine("v0.0.2:");
Console.WriteLine("Blue screen with cursor. not clearing.");
Console.WriteLine("v0.0.3:");
Console.WriteLine("Command Line Shell with Broken CLI, but working unresponsive GUI, but with DOS-like Shell. I/O, Audio, Multithreading, Shutdown doesn't work.");
}
else if (comd == "help")
{
Console.WriteLine("Copyright 2015 Thontelix TriangleOS. Special thanks to Cosmos .net ASM Compiler.");
Console.WriteLine("type CHANGELOG to get version changes.");
Console.WriteLine("How to use:");
Console.WriteLine("After every typed command, press enter. the output of command will be detailed. If there is blinking bottom line cursor, then you need to input something, if it doesn't give any feedback, then its operating a activity. If you want GUI, type 'boot' and press enter.");
Console.WriteLine("Commands:");
Console.WriteLine("shutdown - Gives you ability to safe turn off PC");
Console.WriteLine("boot - Boots you into TriangleOS");
Console.WriteLine("reboot - Reboots your PC");
Console.WriteLine("cd - Travel through directories");
Console.WriteLine("dir - Read content of directory");
Console.WriteLine("echo - Backs string you enter");
Console.WriteLine("shutdown - Backs string you enter");
Console.WriteLine("soundtest - Speakers Drivers. Doesnt work for now");
Console.WriteLine("help - Gives you list of commands.");
Console.WriteLine("paint - Console Paint App. CAUTION:Not for epilepsy persons.");
}
else
{
Console.WriteLine("That command doesn't exist. :(");
}
}
``` | priority | dup tries to pop more stuff from analytical stack than there is log error exception system exception error compiling method systemvoidkernelcommandsinputcommand system exception opcode il dup tries to pop more stuff from analytical stack than there is at cosmos ilopcode interpretstacktypes idictionary aopcodes stack astack boolean asituationchanged amaxrecursiondepth in c users luka desktop cosmos master source cosmos ilopcode cs line at cosmos appassembler interpretinstructionstodeterminestacktypes list acurrentgroup in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler emitinstructions methodinfo amethod list acurrentgroup boolean in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line system exception opcode il dup tries to pop more stuff from analytical stack than there is at cosmos ilopcode interpretstacktypes idictionary aopcodes stack astack boolean asituationchanged amaxrecursiondepth in c users luka desktop cosmos master source cosmos ilopcode cs line at cosmos appassembler interpretinstructionstodeterminestacktypes list acurrentgroup in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler emitinstructions methodinfo amethod list acurrentgroup boolean in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line end of inner exception stack trace at cosmos appassembler processmethod methodinfo amethod list aopcodes in c users luka desktop cosmos master source cosmos appassembler cs line at cosmos ilscanner assemble in c users luka desktop cosmos master source cosmos ilscanner cs line at cosmos ilscanner execute methodbase astartmethod in c users luka desktop cosmos master source cosmos ilscanner cs line at cosmos compilerengine execute in c users luka desktop cosmos master source cosmos compilerengine cs line and there is code where it heappen c public static void inputcommand console write d command comd console readline comd comd tolower if comd reboot h power restart else if comd shutdown h power shutdown else if comd echo console write echo arg console readline console writeline arg else if comd notepad system cli applications notepad else if comd cls console clear console writeline triangleos console writeline else if comd soundtest console write frequency arg console readline console write duration optarg console readline console write eh this isn t implemted right now h multimedia speakers callsound int parse arg int parse optarg else if comd boot console writeline starting triangleos drivers processmanager process audio new processmanager process processmanager process graphics new processmanager process graphics processthread new system threading thread h graphics lowlevel init processmanager process mouse new processmanager process mouse processthread new system threading thread h mouse initmouse graphics start mouse start audio processthread new system threading thread h multimedia speakers intailizeaudio kernel gui else if comd cliboot system cli controls textbox text new system cli controls textbox text x text y text length text drawtextbox text typeinto system cli controls button ok new system cli controls button ok y ok x ok width ok height ok text ok ok drawbutton else if comd calculator system cli applications calculator else if comd cd h graphics console erro impossible operation performed can t request i o while it isn t running else if comd dir console writeline this isn t folder you can use to go up folder else if comd paint system cli applications paint else if comd changelog console writeline you are running version dev only luka see the dev console writeline console writeline blue screen with cursor not clearing console writeline console writeline command line shell with broken cli but working unresponsive gui but with dos like shell i o audio multithreading shutdown doesn t work else if comd help console writeline copyright thontelix triangleos special thanks to cosmos net asm compiler console writeline type changelog to get version changes console writeline how to use console writeline after every typed command press enter the output of command will be detailed if there is blinking bottom line cursor then you need to input something if it doesn t give any feedback then its operating a activity if you want gui type boot and press enter console writeline commands console writeline shutdown gives you ability to safe turn off pc console writeline boot boots you into triangleos console writeline reboot reboots your pc console writeline cd travel through directories console writeline dir read content of directory console writeline echo backs string you enter console writeline shutdown backs string you enter console writeline soundtest speakers drivers doesnt work for now console writeline help gives you list of commands console writeline paint console paint app caution not for epilepsy persons else console writeline that command doesn t exist | 1 |
240,185 | 7,800,589,404 | IssuesEvent | 2018-06-09 11:19:53 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0010214:
improve calendar performance by reducing the number of recurring events fetched | Calendar Mantis high priority | **Reported by pschuele on 5 Sep 2014 16:16**
**Version:** Collin (2013.10.8crowdfunding2)
improve calendar performance by reducing the number of recurring events fetched
-> because we load all related data for every recurring candidate even if it does not match the period filter
| 1.0 | 0010214:
improve calendar performance by reducing the number of recurring events fetched - **Reported by pschuele on 5 Sep 2014 16:16**
**Version:** Collin (2013.10.8crowdfunding2)
improve calendar performance by reducing the number of recurring events fetched
-> because we load all related data for every recurring candidate even if it does not match the period filter
| priority | improve calendar performance by reducing the number of recurring events fetched reported by pschuele on sep version collin improve calendar performance by reducing the number of recurring events fetched gt because we load all related data for every recurring candidate even if it does not match the period filter | 1 |
82,368 | 3,605,944,718 | IssuesEvent | 2016-02-04 08:57:04 | TrinityCore/TrinityCore | https://api.github.com/repos/TrinityCore/TrinityCore | closed | [3.3.5] [Script] ICC Emblem of Frost bug | Branch-3.3.5a Comp-Core Priority-High | Dead characters get frost without raid save !
How to reproduce:
1- Need 2 character. create a raid and enter into ICC.
2- Before first boss let's one of characters die and realese spirit. don't resurrect.
3- Kill first boss with another character.
4- Two characters get the frost. but the dead character did'nt get save. and can do it again!
This worked for next bosses.
rev: c04c409 | 1.0 | [3.3.5] [Script] ICC Emblem of Frost bug - Dead characters get frost without raid save !
How to reproduce:
1- Need 2 character. create a raid and enter into ICC.
2- Before first boss let's one of characters die and realese spirit. don't resurrect.
3- Kill first boss with another character.
4- Two characters get the frost. but the dead character did'nt get save. and can do it again!
This worked for next bosses.
rev: c04c409 | priority | icc emblem of frost bug dead characters get frost without raid save how to reproduce need character create a raid and enter into icc before first boss let s one of characters die and realese spirit don t resurrect kill first boss with another character two characters get the frost but the dead character did nt get save and can do it again this worked for next bosses rev | 1 |
757,617 | 26,521,577,842 | IssuesEvent | 2023-01-19 03:30:49 | pibolib/hack16-2 | https://api.github.com/repos/pibolib/hack16-2 | closed | Implement enemy dodge system | enhancement gameplay priority:high | This requires a rewrite to the way player bullets are handled in regards to hurting the enemy.
Specifications:
1. Remove direct call to Enemy:take_damage()
2. ~~Create signal on Enemy for bullet_collide~~
**Just calls the collision function directly, no signal needed**
4. ~~On bullet_collide, handle on a case by case (inherited class specific) basis~~
**Handles movement on a case by case basis, dodge() super function to always be called in inherited classes**
6. Always handle dodges before taking damage | 1.0 | Implement enemy dodge system - This requires a rewrite to the way player bullets are handled in regards to hurting the enemy.
Specifications:
1. Remove direct call to Enemy:take_damage()
2. ~~Create signal on Enemy for bullet_collide~~
**Just calls the collision function directly, no signal needed**
4. ~~On bullet_collide, handle on a case by case (inherited class specific) basis~~
**Handles movement on a case by case basis, dodge() super function to always be called in inherited classes**
6. Always handle dodges before taking damage | priority | implement enemy dodge system this requires a rewrite to the way player bullets are handled in regards to hurting the enemy specifications remove direct call to enemy take damage create signal on enemy for bullet collide just calls the collision function directly no signal needed on bullet collide handle on a case by case inherited class specific basis handles movement on a case by case basis dodge super function to always be called in inherited classes always handle dodges before taking damage | 1 |
766,406 | 26,882,646,626 | IssuesEvent | 2023-02-05 20:28:24 | sczerwinski/wavefront-obj-intellij-plugin | https://api.github.com/repos/sczerwinski/wavefront-obj-intellij-plugin | closed | Objects without material (or with non-existing material) not rendered | type:bug resolution:done priority:high component:3d | ### Steps
1. Remove all material references from OBJ file.
OR
2. Remove the material used in an OBJ file.
### Result
The OBJ file is not rendered.
### Expected Result
The OBJ file should be normally rendered, just without a material (default textures/colours).
| 1.0 | Objects without material (or with non-existing material) not rendered - ### Steps
1. Remove all material references from OBJ file.
OR
2. Remove the material used in an OBJ file.
### Result
The OBJ file is not rendered.
### Expected Result
The OBJ file should be normally rendered, just without a material (default textures/colours).
| priority | objects without material or with non existing material not rendered steps remove all material references from obj file or remove the material used in an obj file result the obj file is not rendered expected result the obj file should be normally rendered just without a material default textures colours | 1 |
95,990 | 3,962,917,705 | IssuesEvent | 2016-05-02 18:33:44 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | Support for Esri Feature Service Symbology | addition: feature priority: high | The Esri Feature Service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend/layer selector. Instead, the viewer should interrogate the service and interpret the Esri symbology values to generate an appropriate graphic.
Range, Unique Value and Simple rendering should be supported. | 1.0 | Support for Esri Feature Service Symbology - The Esri Feature Service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend/layer selector. Instead, the viewer should interrogate the service and interpret the Esri symbology values to generate an appropriate graphic.
Range, Unique Value and Simple rendering should be supported. | priority | support for esri feature service symbology the esri feature service does not contain a legend endpoint which is where the viewer will by default attempt to obtain symbology for use in the legend layer selector instead the viewer should interrogate the service and interpret the esri symbology values to generate an appropriate graphic range unique value and simple rendering should be supported | 1 |
80,367 | 3,561,152,316 | IssuesEvent | 2016-01-23 16:13:02 | cuckoosandbox/cuckoo | https://api.github.com/repos/cuckoosandbox/cuckoo | closed | Memory dump only readable by root | Bug (to verify) High Priority | We are running cuckoo from git on ubuntu 14.04, and when a memory dump is taken by libvirt the file is only readable by root, which causes problems for volatility.
Just creating the file before calling libvirt seems to solve the problem.
My ugly hack in cuckoo/lib/common/abstracts.py from line 422:
def dump_memory(self, label, path):
"""Takes a memory dump.
@param path: path to where to store the memory dump.
"""
log.debug("Dumping memory for machine %s", label)
#Creating file before call to libvirt to get read permissions for user after dump
open(path,'w').write('')
conn = self._connect()
try:
self.vms[label].coreDump(path, flags=libvirt.VIR_DUMP_MEMORY_ONLY)
except libvirt.libvirtError as e:
raise CuckooMachineError("Error dumping memory virtual machine "
"{0}: {1}".format(label, e))
finally:
self._disconnect(conn)
| 1.0 | Memory dump only readable by root - We are running cuckoo from git on ubuntu 14.04, and when a memory dump is taken by libvirt the file is only readable by root, which causes problems for volatility.
Just creating the file before calling libvirt seems to solve the problem.
My ugly hack in cuckoo/lib/common/abstracts.py from line 422:
def dump_memory(self, label, path):
"""Takes a memory dump.
@param path: path to where to store the memory dump.
"""
log.debug("Dumping memory for machine %s", label)
#Creating file before call to libvirt to get read permissions for user after dump
open(path,'w').write('')
conn = self._connect()
try:
self.vms[label].coreDump(path, flags=libvirt.VIR_DUMP_MEMORY_ONLY)
except libvirt.libvirtError as e:
raise CuckooMachineError("Error dumping memory virtual machine "
"{0}: {1}".format(label, e))
finally:
self._disconnect(conn)
| priority | memory dump only readable by root we are running cuckoo from git on ubuntu and when a memory dump is taken by libvirt the file is only readable by root which causes problems for volatility just creating the file before calling libvirt seems to solve the problem my ugly hack in cuckoo lib common abstracts py from line def dump memory self label path takes a memory dump param path path to where to store the memory dump log debug dumping memory for machine s label creating file before call to libvirt to get read permissions for user after dump open path w write conn self connect try self vms coredump path flags libvirt vir dump memory only except libvirt libvirterror as e raise cuckoomachineerror error dumping memory virtual machine format label e finally self disconnect conn | 1 |
175,772 | 6,553,922,830 | IssuesEvent | 2017-09-06 01:58:47 | kinueng/bluetooth-scale | https://api.github.com/repos/kinueng/bluetooth-scale | closed | Start scanning after disconnecting from BLE device | enhancement priority/high status/inprogress | The end goal is to allow the app to always be running, but the BLE device can come and go as it pleases. | 1.0 | Start scanning after disconnecting from BLE device - The end goal is to allow the app to always be running, but the BLE device can come and go as it pleases. | priority | start scanning after disconnecting from ble device the end goal is to allow the app to always be running but the ble device can come and go as it pleases | 1 |
73,375 | 3,411,272,070 | IssuesEvent | 2015-12-05 00:57:49 | Ecotrust/COMPASS | https://api.github.com/repos/Ecotrust/COMPASS | closed | info icon on 'Active Tab' layers does nothing | bug High Priority | Either create info-dropdowns (like first tab) or remove icons. | 1.0 | info icon on 'Active Tab' layers does nothing - Either create info-dropdowns (like first tab) or remove icons. | priority | info icon on active tab layers does nothing either create info dropdowns like first tab or remove icons | 1 |
304,248 | 9,329,409,585 | IssuesEvent | 2019-03-28 02:15:41 | InQuest/ThreatKB | https://api.github.com/repos/InQuest/ThreatKB | closed | Dynamically update the title. | high-priority | Let the title be "ThreatKB" everywhere, unless, you're viewing or editing an artifact (C2 IP, C2 DNS, YARA). In that case, make the titles:
* KB: $IP
* KB: $Domain
* KB: YARA Signature Name
Will help identify the appropriate tabs at a glance, when we have multiple open. | 1.0 | Dynamically update the title. - Let the title be "ThreatKB" everywhere, unless, you're viewing or editing an artifact (C2 IP, C2 DNS, YARA). In that case, make the titles:
* KB: $IP
* KB: $Domain
* KB: YARA Signature Name
Will help identify the appropriate tabs at a glance, when we have multiple open. | priority | dynamically update the title let the title be threatkb everywhere unless you re viewing or editing an artifact ip dns yara in that case make the titles kb ip kb domain kb yara signature name will help identify the appropriate tabs at a glance when we have multiple open | 1 |
793,110 | 27,983,086,942 | IssuesEvent | 2023-03-26 11:46:15 | AY2223S2-CS2103T-W12-1/tp | https://api.github.com/repos/AY2223S2-CS2103T-W12-1/tp | closed | Update isSamePerson check to check for equality of nric only | type.Bug priority.High severity.Low | Currently, isSamePerson checks for equality of name and nric.
This may not match expectations as nric should uniquely identify a person and each person should only have one main name. | 1.0 | Update isSamePerson check to check for equality of nric only - Currently, isSamePerson checks for equality of name and nric.
This may not match expectations as nric should uniquely identify a person and each person should only have one main name. | priority | update issameperson check to check for equality of nric only currently issameperson checks for equality of name and nric this may not match expectations as nric should uniquely identify a person and each person should only have one main name | 1 |
530,370 | 15,421,828,030 | IssuesEvent | 2021-03-05 13:36:29 | Systems-Learning-and-Development-Lab/MMM | https://api.github.com/repos/Systems-Learning-and-Development-Lab/MMM | closed | Remove existing ball traces if balls are erased | enhancement priority-high | If user clicks on buttons in the UI to remove balls, then traces should be removed as well.

In response to [comment](https://github.com/Systems-Learning-and-Development-Lab/MMM/issues/40#issuecomment-789879108) | 1.0 | Remove existing ball traces if balls are erased - If user clicks on buttons in the UI to remove balls, then traces should be removed as well.

In response to [comment](https://github.com/Systems-Learning-and-Development-Lab/MMM/issues/40#issuecomment-789879108) | priority | remove existing ball traces if balls are erased if user clicks on buttons in the ui to remove balls then traces should be removed as well in response to | 1 |
230,200 | 7,605,351,971 | IssuesEvent | 2018-04-30 08:35:34 | esaude/esaude-emr-poc | https://api.github.com/repos/esaude/esaude-emr-poc | closed | [Access Control] Cannot login with a user with receptionist (admin) role only | High Priority bug | Actual Results
--
The system is not permitting the login of a user with receptionist (admin) role only
Expected results
--
Should allow the login for all the POC roles
Steps to reproduce
--
In OpenMRS admin page create a user with POC: Receptionist (Admin) role only:
Screenshot/Attachment (Optional)
--
A visual description of the unexpected behaviour.

| 1.0 | [Access Control] Cannot login with a user with receptionist (admin) role only - Actual Results
--
The system is not permitting the login of a user with receptionist (admin) role only
Expected results
--
Should allow the login for all the POC roles
Steps to reproduce
--
In OpenMRS admin page create a user with POC: Receptionist (Admin) role only:
Screenshot/Attachment (Optional)
--
A visual description of the unexpected behaviour.

| priority | cannot login with a user with receptionist admin role only actual results the system is not permitting the login of a user with receptionist admin role only expected results should allow the login for all the poc roles steps to reproduce in openmrs admin page create a user with poc receptionist admin role only screenshot attachment optional a visual description of the unexpected behaviour | 1 |
82,151 | 3,603,541,996 | IssuesEvent | 2016-02-03 19:26:57 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | ListMixin map/extend generic method comments | analyzer-strong-mode area-analyzer priority-high | Right now they don't match the generic method comment signatures on Iterable, so this is an error:
````dart
class _FbList<E> extends Object with ListMixin<E> implements List<E>
```
Marked high-pri because this is easy to hit.
CC @leafpetersen, as this is related to SDK working with strong mode.
(this is sort of area-sdk, but only affects Analyzer strong mode. Not sure best way to label it.) | 1.0 | ListMixin map/extend generic method comments - Right now they don't match the generic method comment signatures on Iterable, so this is an error:
````dart
class _FbList<E> extends Object with ListMixin<E> implements List<E>
```
Marked high-pri because this is easy to hit.
CC @leafpetersen, as this is related to SDK working with strong mode.
(this is sort of area-sdk, but only affects Analyzer strong mode. Not sure best way to label it.) | priority | listmixin map extend generic method comments right now they don t match the generic method comment signatures on iterable so this is an error dart class fblist extends object with listmixin implements list marked high pri because this is easy to hit cc leafpetersen as this is related to sdk working with strong mode this is sort of area sdk but only affects analyzer strong mode not sure best way to label it | 1 |
690,549 | 23,663,955,685 | IssuesEvent | 2022-08-26 18:35:04 | tellor-io/telliot-feeds | https://api.github.com/repos/tellor-io/telliot-feeds | opened | Fix & add DIVA Protocol integration tests | enhancement priority:high | - finish tests in `tests/integrations/diva_protocol/test_report.py`
- finish tests in `tests/integrations/diva_protocol/test_e2e_create_report_settle_pool.py` | 1.0 | Fix & add DIVA Protocol integration tests - - finish tests in `tests/integrations/diva_protocol/test_report.py`
- finish tests in `tests/integrations/diva_protocol/test_e2e_create_report_settle_pool.py` | priority | fix add diva protocol integration tests finish tests in tests integrations diva protocol test report py finish tests in tests integrations diva protocol test create report settle pool py | 1 |
810,845 | 30,264,331,181 | IssuesEvent | 2023-07-07 10:35:18 | zeek/zeek | https://api.github.com/repos/zeek/zeek | closed | Incomplete Zeek-side type generation for exported Spicy unit with unit switch | Priority: High Area: Spicy | The following test should work, but fails since the Zeek-side field for the conditionally filled field `foo` is not generated.
```ruby
# @TEST-EXEC: spicyz -do test.hlto test.spicy test.evt
# @TEST-EXEC: zeek -Cr ${TRACES}/udp-packet.pcap test.hlto main.zeek >output 2>&1
# @TEST-EXEC: btest-diff output
# @TEST-START-FILE test.spicy
module TEST;
public type Message = unit {
switch (1) {
* -> foo: bytes &eod;
};
};
# @TEST-END-FILE
# @TEST-START-FILE test.evt
import TEST;
protocol analyzer spicy::Test over UDP:
port 0/udp - 42000/udp,
parse with TEST::Message;
export TEST::Message;
on TEST::Message -> event TEST_ZEEK::MessageEvt(self);
# @TEST-END-FILE
# @TEST-START-FILE main.zeek
event TEST_ZEEK::MessageEvt(message: TEST::Message)
{ }
# @TEST-END-FILE
```
As of `6.1.0-dev.152` (ccfc68fccd9b0635210bf951ee6d4769f997d157) this emits the following runtime warning:
```
XXXXXXXXXX.XXXXXX analyzer error in <...>/test.evt, line 8: Event parameter mismatch, no matching record field for field 'foo'
```
This looks similar to #3083 with the additional twist of a unit switch. | 1.0 | Incomplete Zeek-side type generation for exported Spicy unit with unit switch - The following test should work, but fails since the Zeek-side field for the conditionally filled field `foo` is not generated.
```ruby
# @TEST-EXEC: spicyz -do test.hlto test.spicy test.evt
# @TEST-EXEC: zeek -Cr ${TRACES}/udp-packet.pcap test.hlto main.zeek >output 2>&1
# @TEST-EXEC: btest-diff output
# @TEST-START-FILE test.spicy
module TEST;
public type Message = unit {
switch (1) {
* -> foo: bytes &eod;
};
};
# @TEST-END-FILE
# @TEST-START-FILE test.evt
import TEST;
protocol analyzer spicy::Test over UDP:
port 0/udp - 42000/udp,
parse with TEST::Message;
export TEST::Message;
on TEST::Message -> event TEST_ZEEK::MessageEvt(self);
# @TEST-END-FILE
# @TEST-START-FILE main.zeek
event TEST_ZEEK::MessageEvt(message: TEST::Message)
{ }
# @TEST-END-FILE
```
As of `6.1.0-dev.152` (ccfc68fccd9b0635210bf951ee6d4769f997d157) this emits the following runtime warning:
```
XXXXXXXXXX.XXXXXX analyzer error in <...>/test.evt, line 8: Event parameter mismatch, no matching record field for field 'foo'
```
This looks similar to #3083 with the additional twist of a unit switch. | priority | incomplete zeek side type generation for exported spicy unit with unit switch the following test should work but fails since the zeek side field for the conditionally filled field foo is not generated ruby test exec spicyz do test hlto test spicy test evt test exec zeek cr traces udp packet pcap test hlto main zeek output test exec btest diff output test start file test spicy module test public type message unit switch foo bytes eod test end file test start file test evt import test protocol analyzer spicy test over udp port udp udp parse with test message export test message on test message event test zeek messageevt self test end file test start file main zeek event test zeek messageevt message test message test end file as of dev this emits the following runtime warning xxxxxxxxxx xxxxxx analyzer error in test evt line event parameter mismatch no matching record field for field foo this looks similar to with the additional twist of a unit switch | 1 |
190,145 | 6,810,323,568 | IssuesEvent | 2017-11-05 03:59:03 | ballerinalang/composer | https://api.github.com/repos/ballerinalang/composer | closed | [Fork-Join] Cannot add the a fork join | 0.95 Priority/Highest Severity/Blocker | Pack: https://wso2.org/jenkins/view/All%20Builds/job/ballerinalang/job/tools-distribution/880/org.ballerinalang.tools$ballerina-tools/
Cannot add a fork join as shown in the attached image

| 1.0 | [Fork-Join] Cannot add the a fork join - Pack: https://wso2.org/jenkins/view/All%20Builds/job/ballerinalang/job/tools-distribution/880/org.ballerinalang.tools$ballerina-tools/
Cannot add a fork join as shown in the attached image

| priority | cannot add the a fork join pack cannot add a fork join as shown in the attached image | 1 |
134,267 | 5,222,927,919 | IssuesEvent | 2017-01-27 10:18:58 | hpi-swt2/workshop-portal | https://api.github.com/repos/hpi-swt2/workshop-portal | closed | Custom variable application fields in application | High Priority needs acceptance team-helene | - [ ] render fields on application fill-out page
**Dependency:**
#322 | 1.0 | Custom variable application fields in application - - [ ] render fields on application fill-out page
**Dependency:**
#322 | priority | custom variable application fields in application render fields on application fill out page dependency | 1 |
162,886 | 6,177,832,534 | IssuesEvent | 2017-07-02 05:53:22 | redox-os/ion | https://api.github.com/repos/redox-os/ion | closed | Incomplete Signal Handling | enhancement high-priority | - [x] Implement SIGINT Handling
- [x] Implement SIGTERM Handling
- [ ] Ensure that SIGPIPE is working correctly
- [X] Implement SIGSTOP Handling
- [X] Implement SIGCONT Handling
- [x] Implement `bg`
- [ ] Implement `fg`
- [ ] Implement SIGHUP Handling
- [ ] disown
- [ ] suspend | 1.0 | Incomplete Signal Handling - - [x] Implement SIGINT Handling
- [x] Implement SIGTERM Handling
- [ ] Ensure that SIGPIPE is working correctly
- [X] Implement SIGSTOP Handling
- [X] Implement SIGCONT Handling
- [x] Implement `bg`
- [ ] Implement `fg`
- [ ] Implement SIGHUP Handling
- [ ] disown
- [ ] suspend | priority | incomplete signal handling implement sigint handling implement sigterm handling ensure that sigpipe is working correctly implement sigstop handling implement sigcont handling implement bg implement fg implement sighup handling disown suspend | 1 |
651,642 | 21,484,116,610 | IssuesEvent | 2022-04-26 20:56:49 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | [Bug] Crash : Ankidroid crashes on checking media. | Priority-High Bug Reproduced | ###### Reproduction Steps
- Updated from 2.15.6 (Play store ver.) to 2.16alpha58
and when i opened a card i started getting "failed to load : failed to load ...."

- So i decided to use check media which instantly crashed the app.
- After that i decided to force sync from anki web to ankidroid which also crashed the app.
- Decided to check if my library was corrupted for some reason or not by downloading a parallel of 2.15.6 on which it seemed to work perfectly, updated the parallel to alpha58 and it was also crashing.
- The alpha also doesn't seem to be detecting the backups neither is it creating them properly from what i can tell.

The only one which was good on 16/04 was made by the stable parallel.
- Just to confirm if the problem was from this alpha or from before this I decided to test on alpha 56 parallel and got the same issues uninstalled the parallel and installed alpha 55 parallel and there wasn't a crash this time, decided to update 55 to 56 and test again and there wasn't a crash, same with from 56 to 58.
- ~~Btw I did try to reinstall main alpha58 directly when i was originally getting the crash before the above point happened and the crashes still happened.~~ I thought i did but seems like i hadn't installed it directly and when I just did it seems to work fine on installing directly and this only happens if you update directly from 2.15.6 to 2.16 Alpha56/58/59.
- Have confirmed that it doesn't crash anymore on main after uninstalling 58 and installing 55 then 56 and 58 again after that.
- ~~Backup still don't seem to be being created properly but am not sure yet, so will confirm after I am.~~ It is detecting backups again and the backups are being created properly.
- Also half of this happened while writing this so it might seem like a story.
###### Expected Result
App checks media , directly syncs from anki web to ankidroid , Backup being created properly.
###### Actual Result
App crashes if you have updated directly to alpha 58 or installing alpha58 directly over current library.
###### Debug info
AnkiDroid Version = 2.16alpha58
Android Version = 11
Manufacturer = Xiaomi
Model = Redmi Note 9 Pro Max
Hardware = qcom
Webview User Agent = Mozilla/5.0 (Linux; Android 11; Redmi Note 9 Pro Max Build/RKQ1.200826.002; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/100.0.4896.88 Mobile Safari/537.36
ACRA UUID = d9199133-a62c-4572-ad4f-899c4d0b5307
Scheduler = std2
Crash Reports Enabled = true
DatabaseV2 Enabled = true
###### Research
*Enter an [x] character to confirm the points below:*
- [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
- [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
- [x] I have searched for similar existing issues here and on the user forum
- [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
| 1.0 | [Bug] Crash : Ankidroid crashes on checking media. - ###### Reproduction Steps
- Updated from 2.15.6 (Play store ver.) to 2.16alpha58
and when i opened a card i started getting "failed to load : failed to load ...."

- So i decided to use check media which instantly crashed the app.
- After that i decided to force sync from anki web to ankidroid which also crashed the app.
- Decided to check if my library was corrupted for some reason or not by downloading a parallel of 2.15.6 on which it seemed to work perfectly, updated the parallel to alpha58 and it was also crashing.
- The alpha also doesn't seem to be detecting the backups neither is it creating them properly from what i can tell.

The only one which was good on 16/04 was made by the stable parallel.
- Just to confirm if the problem was from this alpha or from before this I decided to test on alpha 56 parallel and got the same issues uninstalled the parallel and installed alpha 55 parallel and there wasn't a crash this time, decided to update 55 to 56 and test again and there wasn't a crash, same with from 56 to 58.
- ~~Btw I did try to reinstall main alpha58 directly when i was originally getting the crash before the above point happened and the crashes still happened.~~ I thought i did but seems like i hadn't installed it directly and when I just did it seems to work fine on installing directly and this only happens if you update directly from 2.15.6 to 2.16 Alpha56/58/59.
- Have confirmed that it doesn't crash anymore on main after uninstalling 58 and installing 55 then 56 and 58 again after that.
- ~~Backup still don't seem to be being created properly but am not sure yet, so will confirm after I am.~~ It is detecting backups again and the backups are being created properly.
- Also half of this happened while writing this so it might seem like a story.
###### Expected Result
App checks media , directly syncs from anki web to ankidroid , Backup being created properly.
###### Actual Result
App crashes if you have updated directly to alpha 58 or installing alpha58 directly over current library.
###### Debug info
AnkiDroid Version = 2.16alpha58
Android Version = 11
Manufacturer = Xiaomi
Model = Redmi Note 9 Pro Max
Hardware = qcom
Webview User Agent = Mozilla/5.0 (Linux; Android 11; Redmi Note 9 Pro Max Build/RKQ1.200826.002; wv) AppleWebKit/537.36 (KHTML, like Gecko) Version/4.0 Chrome/100.0.4896.88 Mobile Safari/537.36
ACRA UUID = d9199133-a62c-4572-ad4f-899c4d0b5307
Scheduler = std2
Crash Reports Enabled = true
DatabaseV2 Enabled = true
###### Research
*Enter an [x] character to confirm the points below:*
- [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
- [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
- [x] I have searched for similar existing issues here and on the user forum
- [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
| priority | crash ankidroid crashes on checking media reproduction steps updated from play store ver to and when i opened a card i started getting failed to load failed to load so i decided to use check media which instantly crashed the app after that i decided to force sync from anki web to ankidroid which also crashed the app decided to check if my library was corrupted for some reason or not by downloading a parallel of on which it seemed to work perfectly updated the parallel to and it was also crashing the alpha also doesn t seem to be detecting the backups neither is it creating them properly from what i can tell the only one which was good on was made by the stable parallel just to confirm if the problem was from this alpha or from before this i decided to test on alpha parallel and got the same issues uninstalled the parallel and installed alpha parallel and there wasn t a crash this time decided to update to and test again and there wasn t a crash same with from to btw i did try to reinstall main directly when i was originally getting the crash before the above point happened and the crashes still happened i thought i did but seems like i hadn t installed it directly and when i just did it seems to work fine on installing directly and this only happens if you update directly from to have confirmed that it doesn t crash anymore on main after uninstalling and installing then and again after that backup still don t seem to be being created properly but am not sure yet so will confirm after i am it is detecting backups again and the backups are being created properly also half of this happened while writing this so it might seem like a story expected result app checks media directly syncs from anki web to ankidroid backup being created properly actual result app crashes if you have updated directly to alpha or installing directly over current library debug info ankidroid version android version manufacturer xiaomi model redmi note pro max hardware qcom webview user agent mozilla linux android redmi note pro max build wv applewebkit khtml like gecko version chrome mobile safari acra uuid scheduler crash reports enabled true enabled true research enter an character to confirm the points below i have read the and am reporting a bug or enhancement request specific to ankidroid i have checked the and the and could not find a solution to my issue i have searched for similar existing issues here and on the user forum optional i have confirmed the issue is not resolved in the latest alpha release | 1 |
468,035 | 13,460,220,533 | IssuesEvent | 2020-09-09 13:20:52 | UniversityOfHelsinkiCS/lomake | https://api.github.com/repos/UniversityOfHelsinkiCS/lomake | closed | Create downloadable csv export for system wide answers | enhancement high priority | Format should be something like this:
| Programme name| Answer 1 | Answer 2|
|----------|:-------------:|------:|
| prog 1 | answer 1 | answer 2 |
| prog 2 | answer 1 | answer 2 |
| prog 3 | answer 1 | answer 2 | | 1.0 | Create downloadable csv export for system wide answers - Format should be something like this:
| Programme name| Answer 1 | Answer 2|
|----------|:-------------:|------:|
| prog 1 | answer 1 | answer 2 |
| prog 2 | answer 1 | answer 2 |
| prog 3 | answer 1 | answer 2 | | priority | create downloadable csv export for system wide answers format should be something like this programme name answer answer prog answer answer prog answer answer prog answer answer | 1 |
380,656 | 11,269,287,171 | IssuesEvent | 2020-01-14 08:34:36 | wazuh/wazuh-kibana-app | https://api.github.com/repos/wazuh/wazuh-kibana-app | closed | Wazuh FIM - "show files" timeout when monitor lost of files | bug help wanted priority/high | | Wazuh | Elastic | Rev |
| ----- | ------- | --- |
| 3.9.2 | 7.1.1 | 444 |
**Description**
I cannot use show files when lots of files is monitored with FIM.
I increased the timeout in the kibana configs, but i got timeout after 20 sec, or 'Maximum call stack size exceeded'.
My question:
- how i increase the API timeout?
**Steps to reproduce**
1. Go to 'Integrity monitoring'
2. Click on 'show files'
3. see error 'Maximum call stack size exceeded'
4. search with "wp"
5. see the file list
6. click on "Last modified ^ "
7. see error "Wazuh API error: 3021 - Timeout executing API request"
**Screenshots**


**Additional context**
I use docker to run manager
kibana.yml :
- elasticsearch.requestTimeout: 900000
kibana wazuh plugin:
- timout: 90000
the request time after 20sec
I monitor several WP directory.
I don't monitor the cache files.
| 1.0 | Wazuh FIM - "show files" timeout when monitor lost of files - | Wazuh | Elastic | Rev |
| ----- | ------- | --- |
| 3.9.2 | 7.1.1 | 444 |
**Description**
I cannot use show files when lots of files is monitored with FIM.
I increased the timeout in the kibana configs, but i got timeout after 20 sec, or 'Maximum call stack size exceeded'.
My question:
- how i increase the API timeout?
**Steps to reproduce**
1. Go to 'Integrity monitoring'
2. Click on 'show files'
3. see error 'Maximum call stack size exceeded'
4. search with "wp"
5. see the file list
6. click on "Last modified ^ "
7. see error "Wazuh API error: 3021 - Timeout executing API request"
**Screenshots**


**Additional context**
I use docker to run manager
kibana.yml :
- elasticsearch.requestTimeout: 900000
kibana wazuh plugin:
- timout: 90000
the request time after 20sec
I monitor several WP directory.
I don't monitor the cache files.
| priority | wazuh fim show files timeout when monitor lost of files wazuh elastic rev description i cannot use show files when lots of files is monitored with fim i increased the timeout in the kibana configs but i got timeout after sec or maximum call stack size exceeded my question how i increase the api timeout steps to reproduce go to integrity monitoring click on show files see error maximum call stack size exceeded search with wp see the file list click on last modified see error wazuh api error timeout executing api request screenshots additional context i use docker to run manager kibana yml elasticsearch requesttimeout kibana wazuh plugin timout the request time after i monitor several wp directory i don t monitor the cache files | 1 |
658,418 | 21,892,053,182 | IssuesEvent | 2022-05-20 03:30:36 | TeamSparker/Spark-Server | https://api.github.com/repos/TeamSparker/Spark-Server | opened | [FIX] recordDB.getFeedRecordsByRoomIds UTC + 9 처리 | 🦋 영권 High-Priority 🚨 🚨 ISSUE 🚨 | ## 📌 Issue
- CURRENT_DATE - INTERVAL '7 days' 가 UTC + 9 기준이 아님
## 📝 To-do
- [ ] Fix it
| 1.0 | [FIX] recordDB.getFeedRecordsByRoomIds UTC + 9 처리 - ## 📌 Issue
- CURRENT_DATE - INTERVAL '7 days' 가 UTC + 9 기준이 아님
## 📝 To-do
- [ ] Fix it
| priority | recorddb getfeedrecordsbyroomids utc 처리 📌 issue current date interval days 가 utc 기준이 아님 📝 to do fix it | 1 |
804,757 | 29,499,641,081 | IssuesEvent | 2023-06-02 20:17:31 | culibraries/library-reporting | https://api.github.com/repos/culibraries/library-reporting | closed | Trace Lost Reports | High Priority | Reports that are ran using a combination of status and item notes to check for missing items before they are declared missing.
Items to Export: Title List
Filter
- [ ] Note field = "xxx"
Sierra Query:
ITEM STATUS equal to "l" AND (ITEM LOCATION equal to "norst" OR ITEM LOCATION equal to "norov" OR ITEM LOCATION equal to "asnst" OR ITEM LOCATION equal to "asnov" OR ITEM LOCATION equal to "artst" OR ITEM LOCATION equal to "artov" OR ITEM LOCATION equal to "scist" OR ITEM LOCATION equal to "sciov") AND ITEM NOTE All Fields don't have "march 2022" | 1.0 | Trace Lost Reports - Reports that are ran using a combination of status and item notes to check for missing items before they are declared missing.
Items to Export: Title List
Filter
- [ ] Note field = "xxx"
Sierra Query:
ITEM STATUS equal to "l" AND (ITEM LOCATION equal to "norst" OR ITEM LOCATION equal to "norov" OR ITEM LOCATION equal to "asnst" OR ITEM LOCATION equal to "asnov" OR ITEM LOCATION equal to "artst" OR ITEM LOCATION equal to "artov" OR ITEM LOCATION equal to "scist" OR ITEM LOCATION equal to "sciov") AND ITEM NOTE All Fields don't have "march 2022" | priority | trace lost reports reports that are ran using a combination of status and item notes to check for missing items before they are declared missing items to export title list filter note field xxx sierra query item status equal to l and item location equal to norst or item location equal to norov or item location equal to asnst or item location equal to asnov or item location equal to artst or item location equal to artov or item location equal to scist or item location equal to sciov and item note all fields don t have march | 1 |
577,199 | 17,105,124,842 | IssuesEvent | 2021-07-09 16:29:43 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Segmentation fault when a Tensor backward hook removes itself | high priority module: autograd module: crash triaged | ## 🐛 Bug
If a Tensor backward hook removes itself during execution *and* returns a modified gradient, PyTorch will segfault.
## To Reproduce
```python
import torch
def install_hook(tensor):
handle = None
def hook(tensor):
handle.remove()
return torch.zeros_like(tensor)
handle = tensor.register_hook(hook)
def test_hook():
t = torch.ones((1, 5))
t.requires_grad_()
install_hook(t)
(t ** 2).mean().backward()
print(t.grad)
if __name__ == '__main__':
test_hook()
```
```
$ python test_hook.py
Segmentation fault (core dumped)
```
Stacktrace, as recorded by gdb:
```
#0 func_get_name (op=0x7fffb186b790, _unused_ignored=0x0) at /tmp/build/80754af9/python_1599203911753/work/Objects/funcobject.c:299
#1 0x0000555555692f65 in _PyObject_GenericGetAttrWithDict (obj=0x7fffb186b790, name=0x7ffff74b1770, dict=0x0, suppress=0) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:1254
#2 0x00005555556a90fb in PyObject_GetAttrString (v=0x7fffb186b790, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:846
#3 0x00005555555e5beb in PyObject_HasAttrString (v=<optimized out>, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:854
#4 0x00007fffdc92377b in hook_name(_object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#5 0x00007fffdc923dcc in torch::autograd::PyFunctionPreHook::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#6 0x00007fffd60cb1ff in torch::autograd::AccumulateGrad::apply(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#7 0x00007fffd60c651b in torch::autograd::Node::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#8 0x00007fffd60c1dc3 in torch::autograd::Engine::evaluate_function(std::shared_ptr<torch::autograd::GraphTask>&, torch::autograd::Node*, torch::autograd::InputBuffer&, std::shared_ptr<torch::autograd::ReadyQueue> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#9 0x00007fffd60c2698 in torch::autograd::Engine::thread_main(std::shared_ptr<torch::autograd::GraphTask> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#10 0x00007fffd60bd481 in torch::autograd::Engine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#11 0x00007fffdc9177c1 in torch::autograd::python::PythonEngine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#12 0x00007fffd60bf375 in torch::autograd::Engine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#13 0x00007fffdc917711 in torch::autograd::python::PythonEngine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#14 0x00007fffdc91818c in THPEngine_run_backward(_object*, _object*, _object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#15 0x00005555556d3f76 in cfunction_call_varargs (kwargs=<optimized out>, args=<optimized out>, func=0x7ffff74b58b0) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:742
#16 PyCFunction_Call (func=0x7ffff74b58b0, args=<optimized out>, kwargs=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:772
#17 0x000055555569185f in _PyObject_MakeTpCall (callable=0x7ffff74b58b0, args=<optimized out>, nargs=<optimized out>, keywords=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:159
#18 0x0000555555715b90 in _PyObject_Vectorcall (kwnames=0x7fffb6e94400, nargsf=<optimized out>, args=<optimized out>, callable=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:125
#19 call_function (kwnames=0x7fffb6e94400, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#20 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515
#21 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb2384660, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x7fffb6e2de98, kwargs=0x7ffff75807c8, kwcount=<optimized out>, kwstep=1, defs=0x7fffb23883c8, defcount=5, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7ffff74ae7f0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#22 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75807a8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435
#23 0x000055555565411a in _PyObject_Vectorcall (kwnames=0x7fffb6e2de80, nargsf=<optimized out>, args=<optimized out>, callable=0x7fffb2355ca0) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#24 call_function (kwnames=0x7fffb6e2de80, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#25 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515
#26 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb6e35450, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x7ffff752e3c0, kwcount=<optimized out>, kwstep=1, defs=0x7fffb6e39aa8, defcount=4, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7fffb6f83970) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#27 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff752e3b8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435
#28 0x000055555565475e in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff752e3b8, callable=0x7fffb6d64790) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#29 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#30 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3486
#31 0x00005555556df86b in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:283
#32 _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75bc5b0, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:410
#33 0x0000555555652b84 in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff75bc5b0, callable=0x7fffb18c3c10) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#34 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#35 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3500
#36 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7ffff74943a0, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x0, kwcount=<optimized out>, kwstep=2, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#37 0x00005555556df754 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4327
#38 0x000055555576dedc in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:718
#39 0x000055555576df84 in run_eval_code_obj (co=0x7ffff74943a0, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1125
#40 0x00005555557a01f4 in run_mod (mod=<optimized out>, filename=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, flags=<optimized out>, arena=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1147
#41 0x00005555556686e1 in PyRun_FileExFlags (fp=0x5555558eaf90, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1063
#42 0x0000555555668ac6 in PyRun_SimpleFileExFlags (fp=0x5555558eaf90, filename=<optimized out>, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:428
#43 0x000055555566998b in pymain_run_file (cf=0x7fffffffdbd8, config=0x5555558eb5b0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:387
#44 pymain_run_python (exitcode=0x7fffffffdbd0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:612
#45 Py_RunMain () at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:691
#46 0x00005555557a2d19 in Py_BytesMain (argc=<optimized out>, argv=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:1137
#47 0x00007ffff7db00b3 in __libc_start_main (main=0x55555566a460 <main>, argc=2, argv=0x7fffffffddd8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffddc8) at ../csu/libc-start.c:308
#48 0x0000555555732e93 in _start () at ../sysdeps/x86_64/elf/start.S:103
```
## Expected behavior
It doesn't segfault. If a hook removing itself is considered an error, it should raise an exception.
## Environment
```
PyTorch version: 1.9.0a0+git6af4e99
Is debug build: False
CUDA used to build PyTorch: 10.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.2 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: 13.0.0-++20210418052640+d480f968ad8b-1~exp1~20210418153358.383
CMake version: version 3.18.2
Python version: 3.8 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: 10.1.243
GPU models and configuration: GPU 0: GeForce RTX 2080 SUPER
Nvidia driver version: 450.119.03
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.2
[pip3] numpydoc==1.1.0
[pip3] torch==1.9.0a0+git9def776
[conda] blas 1.0 mkl
[conda] mkl 2020.2 256
[conda] mkl-include 2021.2.0 h06a4308_296
[conda] mkl-service 2.3.0 py38he904b0f_0
[conda] mkl_fft 1.2.0 py38h23d657b_0
[conda] mkl_random 1.1.1 py38h0573a6f_0
[conda] numpy 1.19.2 py38h54aff64_0
[conda] numpy-base 1.19.2 py38hfa32c7d_0
[conda] numpydoc 1.1.0 pyhd3eb1b0_1
[conda] torch 1.9.0a0+git9def776 pypi_0 pypi
```
## Additional context
Tested on PyTorch built from master as of 9def776cd6d775779db9839d942a01b4b082175c as well as on 1.8.1
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @gqchen @pearu @nikitaved @soulitzer @Lezcano | 1.0 | Segmentation fault when a Tensor backward hook removes itself - ## 🐛 Bug
If a Tensor backward hook removes itself during execution *and* returns a modified gradient, PyTorch will segfault.
## To Reproduce
```python
import torch
def install_hook(tensor):
handle = None
def hook(tensor):
handle.remove()
return torch.zeros_like(tensor)
handle = tensor.register_hook(hook)
def test_hook():
t = torch.ones((1, 5))
t.requires_grad_()
install_hook(t)
(t ** 2).mean().backward()
print(t.grad)
if __name__ == '__main__':
test_hook()
```
```
$ python test_hook.py
Segmentation fault (core dumped)
```
Stacktrace, as recorded by gdb:
```
#0 func_get_name (op=0x7fffb186b790, _unused_ignored=0x0) at /tmp/build/80754af9/python_1599203911753/work/Objects/funcobject.c:299
#1 0x0000555555692f65 in _PyObject_GenericGetAttrWithDict (obj=0x7fffb186b790, name=0x7ffff74b1770, dict=0x0, suppress=0) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:1254
#2 0x00005555556a90fb in PyObject_GetAttrString (v=0x7fffb186b790, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:846
#3 0x00005555555e5beb in PyObject_HasAttrString (v=<optimized out>, name=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/object.c:854
#4 0x00007fffdc92377b in hook_name(_object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#5 0x00007fffdc923dcc in torch::autograd::PyFunctionPreHook::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#6 0x00007fffd60cb1ff in torch::autograd::AccumulateGrad::apply(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#7 0x00007fffd60c651b in torch::autograd::Node::operator()(std::vector<at::Tensor, std::allocator<at::Tensor> >&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#8 0x00007fffd60c1dc3 in torch::autograd::Engine::evaluate_function(std::shared_ptr<torch::autograd::GraphTask>&, torch::autograd::Node*, torch::autograd::InputBuffer&, std::shared_ptr<torch::autograd::ReadyQueue> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#9 0x00007fffd60c2698 in torch::autograd::Engine::thread_main(std::shared_ptr<torch::autograd::GraphTask> const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#10 0x00007fffd60bd481 in torch::autograd::Engine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#11 0x00007fffdc9177c1 in torch::autograd::python::PythonEngine::execute_with_graph_task(std::shared_ptr<torch::autograd::GraphTask> const&, std::shared_ptr<torch::autograd::Node>, torch::autograd::InputBuffer&&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#12 0x00007fffd60bf375 in torch::autograd::Engine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so
#13 0x00007fffdc917711 in torch::autograd::python::PythonEngine::execute(std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&, std::vector<at::Tensor, std::allocator<at::Tensor> > const&, bool, bool, bool, std::vector<torch::autograd::Edge, std::allocator<torch::autograd::Edge> > const&) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#14 0x00007fffdc91818c in THPEngine_run_backward(_object*, _object*, _object*) () from /home/nelhage/sw/anaconda3/lib/python3.8/site-packages/torch/lib/libtorch_python.so
#15 0x00005555556d3f76 in cfunction_call_varargs (kwargs=<optimized out>, args=<optimized out>, func=0x7ffff74b58b0) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:742
#16 PyCFunction_Call (func=0x7ffff74b58b0, args=<optimized out>, kwargs=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:772
#17 0x000055555569185f in _PyObject_MakeTpCall (callable=0x7ffff74b58b0, args=<optimized out>, nargs=<optimized out>, keywords=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:159
#18 0x0000555555715b90 in _PyObject_Vectorcall (kwnames=0x7fffb6e94400, nargsf=<optimized out>, args=<optimized out>, callable=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:125
#19 call_function (kwnames=0x7fffb6e94400, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#20 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515
#21 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb2384660, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x7fffb6e2de98, kwargs=0x7ffff75807c8, kwcount=<optimized out>, kwstep=1, defs=0x7fffb23883c8, defcount=5, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7ffff74ae7f0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#22 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75807a8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435
#23 0x000055555565411a in _PyObject_Vectorcall (kwnames=0x7fffb6e2de80, nargsf=<optimized out>, args=<optimized out>, callable=0x7fffb2355ca0) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#24 call_function (kwnames=0x7fffb6e2de80, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#25 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3515
#26 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7fffb6e35450, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x7ffff752e3c0, kwcount=<optimized out>, kwstep=1, defs=0x7fffb6e39aa8, defcount=4, kwdefs=0x0, closure=0x0, name=0x7ffff74ae7f0, qualname=0x7fffb6f83970) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#27 0x00005555556df943 in _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff752e3b8, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:435
#28 0x000055555565475e in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff752e3b8, callable=0x7fffb6d64790) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#29 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#30 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3486
#31 0x00005555556df86b in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:283
#32 _PyFunction_Vectorcall (func=<optimized out>, stack=0x7ffff75bc5b0, nargsf=<optimized out>, kwnames=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Objects/call.c:410
#33 0x0000555555652b84 in _PyObject_Vectorcall (kwnames=0x0, nargsf=<optimized out>, args=0x7ffff75bc5b0, callable=0x7fffb18c3c10) at /tmp/build/80754af9/python_1599203911753/work/Include/cpython/abstract.h:127
#34 call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>, tstate=0x5555558ea520) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4963
#35 _PyEval_EvalFrameDefault (f=<optimized out>, throwflag=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:3500
#36 0x00005555556dea92 in _PyEval_EvalCodeWithName (_co=0x7ffff74943a0, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kwnames=0x0, kwargs=0x0, kwcount=<optimized out>, kwstep=2, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4298
#37 0x00005555556df754 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:4327
#38 0x000055555576dedc in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/ceval.c:718
#39 0x000055555576df84 in run_eval_code_obj (co=0x7ffff74943a0, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1125
#40 0x00005555557a01f4 in run_mod (mod=<optimized out>, filename=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, flags=<optimized out>, arena=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1147
#41 0x00005555556686e1 in PyRun_FileExFlags (fp=0x5555558eaf90, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff761f2c0, locals=0x7ffff761f2c0, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:1063
#42 0x0000555555668ac6 in PyRun_SimpleFileExFlags (fp=0x5555558eaf90, filename=<optimized out>, closeit=1, flags=0x7fffffffdbd8) at /tmp/build/80754af9/python_1599203911753/work/Python/pythonrun.c:428
#43 0x000055555566998b in pymain_run_file (cf=0x7fffffffdbd8, config=0x5555558eb5b0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:387
#44 pymain_run_python (exitcode=0x7fffffffdbd0) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:612
#45 Py_RunMain () at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:691
#46 0x00005555557a2d19 in Py_BytesMain (argc=<optimized out>, argv=<optimized out>) at /tmp/build/80754af9/python_1599203911753/work/Modules/main.c:1137
#47 0x00007ffff7db00b3 in __libc_start_main (main=0x55555566a460 <main>, argc=2, argv=0x7fffffffddd8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffddc8) at ../csu/libc-start.c:308
#48 0x0000555555732e93 in _start () at ../sysdeps/x86_64/elf/start.S:103
```
## Expected behavior
It doesn't segfault. If a hook removing itself is considered an error, it should raise an exception.
## Environment
```
PyTorch version: 1.9.0a0+git6af4e99
Is debug build: False
CUDA used to build PyTorch: 10.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.2 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: 13.0.0-++20210418052640+d480f968ad8b-1~exp1~20210418153358.383
CMake version: version 3.18.2
Python version: 3.8 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: 10.1.243
GPU models and configuration: GPU 0: GeForce RTX 2080 SUPER
Nvidia driver version: 450.119.03
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.2
[pip3] numpydoc==1.1.0
[pip3] torch==1.9.0a0+git9def776
[conda] blas 1.0 mkl
[conda] mkl 2020.2 256
[conda] mkl-include 2021.2.0 h06a4308_296
[conda] mkl-service 2.3.0 py38he904b0f_0
[conda] mkl_fft 1.2.0 py38h23d657b_0
[conda] mkl_random 1.1.1 py38h0573a6f_0
[conda] numpy 1.19.2 py38h54aff64_0
[conda] numpy-base 1.19.2 py38hfa32c7d_0
[conda] numpydoc 1.1.0 pyhd3eb1b0_1
[conda] torch 1.9.0a0+git9def776 pypi_0 pypi
```
## Additional context
Tested on PyTorch built from master as of 9def776cd6d775779db9839d942a01b4b082175c as well as on 1.8.1
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @gqchen @pearu @nikitaved @soulitzer @Lezcano | priority | segmentation fault when a tensor backward hook removes itself 🐛 bug if a tensor backward hook removes itself during execution and returns a modified gradient pytorch will segfault to reproduce python import torch def install hook tensor handle none def hook tensor handle remove return torch zeros like tensor handle tensor register hook hook def test hook t torch ones t requires grad install hook t t mean backward print t grad if name main test hook python test hook py segmentation fault core dumped stacktrace as recorded by gdb func get name op unused ignored at tmp build python work objects funcobject c in pyobject genericgetattrwithdict obj name dict suppress at tmp build python work objects object c in pyobject getattrstring v name at tmp build python work objects object c in pyobject hasattrstring v name at tmp build python work objects object c in hook name object from home nelhage sw lib site packages torch lib libtorch python so in torch autograd pyfunctionprehook operator std vector const from home nelhage sw lib site packages torch lib libtorch python so in torch autograd accumulategrad apply std vector from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd node operator std vector from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine evaluate function std shared ptr torch autograd node torch autograd inputbuffer std shared ptr const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine thread main std shared ptr const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd engine execute with graph task std shared ptr const std shared ptr torch autograd inputbuffer from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd python pythonengine execute with graph task std shared ptr const std shared ptr torch autograd inputbuffer from home nelhage sw lib site packages torch lib libtorch python so in torch autograd engine execute std vector const std vector const bool bool bool std vector const from home nelhage sw lib site packages torch lib libtorch cpu so in torch autograd python pythonengine execute std vector const std vector const bool bool bool std vector const from home nelhage sw lib site packages torch lib libtorch python so in thpengine run backward object object object from home nelhage sw lib site packages torch lib libtorch python so in cfunction call varargs kwargs args func at tmp build python work objects call c pycfunction call func args kwargs at tmp build python work objects call c in pyobject maketpcall callable args nargs keywords at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in function code fastcall globals nargs args co at tmp build python work objects call c pyfunction vectorcall func stack nargsf kwnames at tmp build python work objects call c in pyobject vectorcall kwnames nargsf args callable at tmp build python work include cpython abstract h call function kwnames oparg pp stack tstate at tmp build python work python ceval c pyeval evalframedefault f throwflag at tmp build python work python ceval c in pyeval evalcodewithname co globals locals args argcount kwnames kwargs kwcount kwstep defs defcount kwdefs closure name qualname at tmp build python work python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at tmp build python work python ceval c in pyeval evalcode co globals locals at tmp build python work python ceval c in run eval code obj co globals locals at tmp build python work python pythonrun c in run mod mod filename globals locals flags arena at tmp build python work python pythonrun c in pyrun fileexflags fp filename str start globals locals closeit flags at tmp build python work python pythonrun c in pyrun simplefileexflags fp filename closeit flags at tmp build python work python pythonrun c in pymain run file cf config at tmp build python work modules main c pymain run python exitcode at tmp build python work modules main c py runmain at tmp build python work modules main c in py bytesmain argc argv at tmp build python work modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start at sysdeps elf start s expected behavior it doesn t segfault if a hook removing itself is considered an error it should raise an exception environment pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version cmake version version python version bit runtime is cuda available true cuda runtime version gpu models and configuration gpu geforce rtx super nvidia driver version cudnn version could not collect hip runtime version n a miopen runtime version n a versions of relevant libraries numpy numpydoc torch blas mkl mkl mkl include mkl service mkl fft mkl random numpy numpy base numpydoc torch pypi pypi additional context tested on pytorch built from master as of as well as on cc ezyang gchanan bdhirsh jbschlosser alband gqchen pearu nikitaved soulitzer lezcano | 1 |
321,998 | 9,811,221,873 | IssuesEvent | 2019-06-12 22:49:27 | keystonejs/keystone-5 | https://api.github.com/repos/keystonejs/keystone-5 | closed | isRequired has no effect | bug graphql high priority schema | ``` js
fields: {
description: {
type: Text,
isRequired: true
}
}
```
Expectation is that it fails if `null` (or `''`, aka empty string) in the administration UI on creation.
Actuality: It creates an item without an error.
**edit:** Unclear if this is meant be enforced by GraphQL, but, I can verify that you can `create*` `isRequired` fields as `null` without error too. `isUnique` **is** enforced by GraphQL. | 1.0 | isRequired has no effect - ``` js
fields: {
description: {
type: Text,
isRequired: true
}
}
```
Expectation is that it fails if `null` (or `''`, aka empty string) in the administration UI on creation.
Actuality: It creates an item without an error.
**edit:** Unclear if this is meant be enforced by GraphQL, but, I can verify that you can `create*` `isRequired` fields as `null` without error too. `isUnique` **is** enforced by GraphQL. | priority | isrequired has no effect js fields description type text isrequired true expectation is that it fails if null or aka empty string in the administration ui on creation actuality it creates an item without an error edit unclear if this is meant be enforced by graphql but i can verify that you can create isrequired fields as null without error too isunique is enforced by graphql | 1 |
674,578 | 23,058,357,111 | IssuesEvent | 2022-07-25 07:38:17 | zitadel/zitadel | https://api.github.com/repos/zitadel/zitadel | closed | [OAuth]: add more claims to introspection response | category: docs type: enhancement lang: go category: backend priority: high | RFC7662 Token Introspection https://datatracker.ietf.org/doc/html/rfc7662#section-2.2 specifies that only claim `active` is required to be returned. Additionally we also return `scope`, `client_id`, `username` and all requested claims according to userinfo response (incl. `sub`, `iss`)
We should also add `token_type`, `exp`, `iat`, `nbf`, `aud`, `jti`
This would also allow libraries to cache the response more efficiently.
Requires changes to the OIDC Library.
**Acceptance criteria**
- [x] the claims above are returned on the response
- [x] documentation | 1.0 | [OAuth]: add more claims to introspection response - RFC7662 Token Introspection https://datatracker.ietf.org/doc/html/rfc7662#section-2.2 specifies that only claim `active` is required to be returned. Additionally we also return `scope`, `client_id`, `username` and all requested claims according to userinfo response (incl. `sub`, `iss`)
We should also add `token_type`, `exp`, `iat`, `nbf`, `aud`, `jti`
This would also allow libraries to cache the response more efficiently.
Requires changes to the OIDC Library.
**Acceptance criteria**
- [x] the claims above are returned on the response
- [x] documentation | priority | add more claims to introspection response token introspection specifies that only claim active is required to be returned additionally we also return scope client id username and all requested claims according to userinfo response incl sub iss we should also add token type exp iat nbf aud jti this would also allow libraries to cache the response more efficiently requires changes to the oidc library acceptance criteria the claims above are returned on the response documentation | 1 |
346,289 | 10,410,355,708 | IssuesEvent | 2019-09-13 11:09:52 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | conan create -> conan upload | complex: medium priority: high stage: queue type: look into | It will create a file with the reference created (or the install reference if specified). It will be a pref (without revision).
Together with #5196 would alleviate nicely the `conan create` -> `conan upload` typical CI flow that now requires painful parsings.
@solvingj
| 1.0 | conan create -> conan upload - It will create a file with the reference created (or the install reference if specified). It will be a pref (without revision).
Together with #5196 would alleviate nicely the `conan create` -> `conan upload` typical CI flow that now requires painful parsings.
@solvingj
| priority | conan create conan upload it will create a file with the reference created or the install reference if specified it will be a pref without revision together with would alleviate nicely the conan create conan upload typical ci flow that now requires painful parsings solvingj | 1 |
448,563 | 12,952,690,279 | IssuesEvent | 2020-07-19 21:31:22 | passwall/passwall-desktop | https://api.github.com/repos/passwall/passwall-desktop | closed | Parse CSV data to array | Priority: High Status: Available Type: Enhancement | **Is your feature request related to a problem? Please describe.**
Now import function is added ass seen in the picture. User should import csv files with this.
<img width="533" alt="Screen Shot 2020-07-15 at 23 03 36" src="https://user-images.githubusercontent.com/3696314/87590503-8d665a00-c6ef-11ea-9bb4-ade42c6682c2.png">
I am going to encrypt and add values to database however I need someone to parse this csv content line by line as array. The import handle function is in **/src/renderer/App.vue** like this:
```js
onImport() {
dialog.showOpenDialog({ properties: ['openFile'] }, files => {
if (files.length === 0) return
try {
const fileContent = fs.readFileSync(files[0]).toString()
// TODO: Parse fileContent (CSV data) to array
console.log(fileContent)
} catch (error) {
console.log(error)
}
})
}
```
**Describe the solution you'd like**
Example CSV content is here
https://github.com/passwall/passwall-server/blob/master/example.csv
Can someone parse and give me the values of CSV lines as array and javascript variables?
Thank you. | 1.0 | Parse CSV data to array - **Is your feature request related to a problem? Please describe.**
Now import function is added ass seen in the picture. User should import csv files with this.
<img width="533" alt="Screen Shot 2020-07-15 at 23 03 36" src="https://user-images.githubusercontent.com/3696314/87590503-8d665a00-c6ef-11ea-9bb4-ade42c6682c2.png">
I am going to encrypt and add values to database however I need someone to parse this csv content line by line as array. The import handle function is in **/src/renderer/App.vue** like this:
```js
onImport() {
dialog.showOpenDialog({ properties: ['openFile'] }, files => {
if (files.length === 0) return
try {
const fileContent = fs.readFileSync(files[0]).toString()
// TODO: Parse fileContent (CSV data) to array
console.log(fileContent)
} catch (error) {
console.log(error)
}
})
}
```
**Describe the solution you'd like**
Example CSV content is here
https://github.com/passwall/passwall-server/blob/master/example.csv
Can someone parse and give me the values of CSV lines as array and javascript variables?
Thank you. | priority | parse csv data to array is your feature request related to a problem please describe now import function is added ass seen in the picture user should import csv files with this img width alt screen shot at src i am going to encrypt and add values to database however i need someone to parse this csv content line by line as array the import handle function is in src renderer app vue like this js onimport dialog showopendialog properties files if files length return try const filecontent fs readfilesync files tostring todo parse filecontent csv data to array console log filecontent catch error console log error describe the solution you d like example csv content is here can someone parse and give me the values of csv lines as array and javascript variables thank you | 1 |
390,006 | 11,520,593,648 | IssuesEvent | 2020-02-14 15:05:31 | openforis/collect-earth-online | https://api.github.com/repos/openforis/collect-earth-online | opened | Allow users to edit their institution's imagery sources (and make sure they appear on the Review Project page) | Feature Request High Priority | There is no way to edit the imagery source. The process of adding WMS services is very error prone (e.g. mistyping the JSON parameters or changing the resolution or format) so you should be able to edit the imagery sources (maybe the base ones like Bing Maps or Digital Globe should be " locked" )
http://i.imgur.com/0jW6gaw.png
When I choose the PNOA as a basemap it works fine assessing plots, but if I go back to the project page (using the Edit button) the basemap parameter appears empty, even though PNOA is in the list of avaliable sources. see images.
http://i.imgur.com/zurwJEF.png
http://i.imgur.com/gPUGd6u.png | 1.0 | Allow users to edit their institution's imagery sources (and make sure they appear on the Review Project page) - There is no way to edit the imagery source. The process of adding WMS services is very error prone (e.g. mistyping the JSON parameters or changing the resolution or format) so you should be able to edit the imagery sources (maybe the base ones like Bing Maps or Digital Globe should be " locked" )
http://i.imgur.com/0jW6gaw.png
When I choose the PNOA as a basemap it works fine assessing plots, but if I go back to the project page (using the Edit button) the basemap parameter appears empty, even though PNOA is in the list of avaliable sources. see images.
http://i.imgur.com/zurwJEF.png
http://i.imgur.com/gPUGd6u.png | priority | allow users to edit their institution s imagery sources and make sure they appear on the review project page there is no way to edit the imagery source the process of adding wms services is very error prone e g mistyping the json parameters or changing the resolution or format so you should be able to edit the imagery sources maybe the base ones like bing maps or digital globe should be locked when i choose the pnoa as a basemap it works fine assessing plots but if i go back to the project page using the edit button the basemap parameter appears empty even though pnoa is in the list of avaliable sources see images | 1 |
463,579 | 13,284,356,671 | IssuesEvent | 2020-08-24 06:06:28 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1731] Crash with Crane | Category: Gameplay Priority: High Status: Fixed | Step to reproduce:
- place a new Crane.
- take non-constucted blocks from stockpile (I took Birch log):

- start to build:

- press LMB - Crash:
```
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Components.CraneToolComponent.Build(Player player, Vector3i min, Vector3i max)</size>
```
[Crane Crash.txt](https://github.com/StrangeLoopGames/EcoIssues/files/5094658/Crane.Crash.txt)
If you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine. | 1.0 | [0.9.0 staging-1731] Crash with Crane - Step to reproduce:
- place a new Crane.
- take non-constucted blocks from stockpile (I took Birch log):

- start to build:

- press LMB - Crash:
```
System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Components.CraneToolComponent.Build(Player player, Vector3i min, Vector3i max)</size>
```
[Crane Crash.txt](https://github.com/StrangeLoopGames/EcoIssues/files/5094658/Crane.Crash.txt)
If you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine. | priority | crash with crane step to reproduce place a new crane take non constucted blocks from stockpile i took birch log start to build press lmb crash system nullreferenceexception object reference not set to an instance of an object at eco gameplay components cranetoolcomponent build player player min max if you first take constructed block with crane and build something and after that take birch log and try to build then all will be fine | 1 |
382,811 | 11,320,176,831 | IssuesEvent | 2020-01-21 02:56:40 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | App crash on creating a non-program requisition | Docs: not needed Effort: small Module: indicators Priority: high Priority: release blocker | ## Describe the bug
App crashes when creating a non-program requisition.
### To reproduce
1. Create a non-program requisition
4. See error
### Expected behaviour
No crash
### Proposed Solution
Think it's the state init method
### Version and device info
- App version: 3.1.1 development
- Tablet model:
- OS version:
### Additional context
<img width="1357" alt="image" src="https://user-images.githubusercontent.com/35858975/72573200-b87f1500-3929-11ea-80df-6d194305a805.png">
| 2.0 | App crash on creating a non-program requisition - ## Describe the bug
App crashes when creating a non-program requisition.
### To reproduce
1. Create a non-program requisition
4. See error
### Expected behaviour
No crash
### Proposed Solution
Think it's the state init method
### Version and device info
- App version: 3.1.1 development
- Tablet model:
- OS version:
### Additional context
<img width="1357" alt="image" src="https://user-images.githubusercontent.com/35858975/72573200-b87f1500-3929-11ea-80df-6d194305a805.png">
| priority | app crash on creating a non program requisition describe the bug app crashes when creating a non program requisition to reproduce create a non program requisition see error expected behaviour no crash proposed solution think it s the state init method version and device info app version development tablet model os version additional context img width alt image src | 1 |
162,350 | 6,151,454,059 | IssuesEvent | 2017-06-28 02:38:25 | mercadopago/px-ios | https://api.github.com/repos/mercadopago/px-ios | closed | Salimos de flujo al volver para atrás de revisa y confirma | Priority: High | Deberiamos volver a la ultima pantalla del flujo
| 1.0 | Salimos de flujo al volver para atrás de revisa y confirma - Deberiamos volver a la ultima pantalla del flujo
| priority | salimos de flujo al volver para atrás de revisa y confirma deberiamos volver a la ultima pantalla del flujo | 1 |
351,273 | 10,514,775,904 | IssuesEvent | 2019-09-28 03:32:43 | webhintio/hint | https://api.github.com/repos/webhintio/hint | closed | [Bug] Need to add instructions when clicking on the extension to let users know they should go to the dev tools | area:browser-extension priority:high type:bug | Clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab. This is how it looks like right now:

| 1.0 | [Bug] Need to add instructions when clicking on the extension to let users know they should go to the dev tools - Clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab. This is how it looks like right now:

| priority | need to add instructions when clicking on the extension to let users know they should go to the dev tools clicking in the button that appears close to the address bar should tell the user to go to the dev tools and click on the hint tab this is how it looks like right now | 1 |
768,966 | 26,989,126,236 | IssuesEvent | 2023-02-09 18:22:14 | repobee/repobee | https://api.github.com/repos/repobee/repobee | closed | Add periodic CI job that ensures the installer can run on the current release | priority:high ci/cd | As noted in #1116, v3.8.0 could not be installed from source (and thus not with the installer). Why the install script is installing from source instead of the wheel on PyPi is a bit of a mystery in itself.
Anyway, there is obviously a need for a CI job that ensures that the installer can run according to the install instructions, so this doesn't happen again. | 1.0 | Add periodic CI job that ensures the installer can run on the current release - As noted in #1116, v3.8.0 could not be installed from source (and thus not with the installer). Why the install script is installing from source instead of the wheel on PyPi is a bit of a mystery in itself.
Anyway, there is obviously a need for a CI job that ensures that the installer can run according to the install instructions, so this doesn't happen again. | priority | add periodic ci job that ensures the installer can run on the current release as noted in could not be installed from source and thus not with the installer why the install script is installing from source instead of the wheel on pypi is a bit of a mystery in itself anyway there is obviously a need for a ci job that ensures that the installer can run according to the install instructions so this doesn t happen again | 1 |
670,794 | 22,704,259,426 | IssuesEvent | 2022-07-05 13:26:04 | skyportal/skyportal | https://api.github.com/repos/skyportal/skyportal | closed | Automating email and SMS notification given specific requirements | high priority | Right now, we can set allowEmailAlerts and allowSMSAlert (in addition to slack notifications). Best I can tell, allowEmailAlerts and allowSMSAlerts do not do anything, and their behavior should be merged with the slack infrastructure (where different types of notifications are allowed). | 1.0 | Automating email and SMS notification given specific requirements - Right now, we can set allowEmailAlerts and allowSMSAlert (in addition to slack notifications). Best I can tell, allowEmailAlerts and allowSMSAlerts do not do anything, and their behavior should be merged with the slack infrastructure (where different types of notifications are allowed). | priority | automating email and sms notification given specific requirements right now we can set allowemailalerts and allowsmsalert in addition to slack notifications best i can tell allowemailalerts and allowsmsalerts do not do anything and their behavior should be merged with the slack infrastructure where different types of notifications are allowed | 1 |
463,763 | 13,300,066,025 | IssuesEvent | 2020-08-25 10:44:20 | GluuFederation/oxAuth | https://api.github.com/repos/GluuFederation/oxAuth | closed | If software_statement is present in dynamic registration request, plain claims are lost | bug high priority | ## Describe the issue
If software_statement is present in dynamic registration request, plain claims are lost.
From RFC 7591:
```Client metadata values MAY also be provided in a software statement, as described in Section 2.3 ... If the server supports software statements, client metadata values conveyed in the software statement MUST take precedence over those conveyed using plain JSON elements.```
## Expected behavior
Plain text claims should be present and take part in registration, however those claims that are present in software statement should take higher priority (override plain claims).
## Actual behavior
Only software statement claims take part in registration. Plain claims are ignored.
| 1.0 | If software_statement is present in dynamic registration request, plain claims are lost - ## Describe the issue
If software_statement is present in dynamic registration request, plain claims are lost.
From RFC 7591:
```Client metadata values MAY also be provided in a software statement, as described in Section 2.3 ... If the server supports software statements, client metadata values conveyed in the software statement MUST take precedence over those conveyed using plain JSON elements.```
## Expected behavior
Plain text claims should be present and take part in registration, however those claims that are present in software statement should take higher priority (override plain claims).
## Actual behavior
Only software statement claims take part in registration. Plain claims are ignored.
| priority | if software statement is present in dynamic registration request plain claims are lost describe the issue if software statement is present in dynamic registration request plain claims are lost from rfc client metadata values may also be provided in a software statement as described in section if the server supports software statements client metadata values conveyed in the software statement must take precedence over those conveyed using plain json elements expected behavior plain text claims should be present and take part in registration however those claims that are present in software statement should take higher priority override plain claims actual behavior only software statement claims take part in registration plain claims are ignored | 1 |
785,958 | 27,629,658,333 | IssuesEvent | 2023-03-10 09:53:11 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Make release compatibility filtering opt-in | type:feature priority-2-high breaking status:in-progress v35 | ### What would you like Renovate to be able to do?
Default to no release filtering (today only python exact versions is supported) so that we can add more advanced filtering in future.
### If you have any ideas on how this should be implemented, please tell us here.
I'd like to make it an enum in case we want to have different degrees of filtering in future.
Feature name `releaseConstraintFiltering`, default value `"none"`.
Alternative value `"strict"`, and put this code behind an `if (config.releaseConstraintFiltering === 'strict')`: https://github.com/renovatebot/renovate/blob/8e8cbba0e62ac5ccac0a24de295a468c5c2101b8/lib/modules/datasource/index.ts#L398-L419
### Is this a feature you are interested in implementing yourself?
Maybe | 1.0 | Make release compatibility filtering opt-in - ### What would you like Renovate to be able to do?
Default to no release filtering (today only python exact versions is supported) so that we can add more advanced filtering in future.
### If you have any ideas on how this should be implemented, please tell us here.
I'd like to make it an enum in case we want to have different degrees of filtering in future.
Feature name `releaseConstraintFiltering`, default value `"none"`.
Alternative value `"strict"`, and put this code behind an `if (config.releaseConstraintFiltering === 'strict')`: https://github.com/renovatebot/renovate/blob/8e8cbba0e62ac5ccac0a24de295a468c5c2101b8/lib/modules/datasource/index.ts#L398-L419
### Is this a feature you are interested in implementing yourself?
Maybe | priority | make release compatibility filtering opt in what would you like renovate to be able to do default to no release filtering today only python exact versions is supported so that we can add more advanced filtering in future if you have any ideas on how this should be implemented please tell us here i d like to make it an enum in case we want to have different degrees of filtering in future feature name releaseconstraintfiltering default value none alternative value strict and put this code behind an if config releaseconstraintfiltering strict is this a feature you are interested in implementing yourself maybe | 1 |
792,441 | 27,960,524,707 | IssuesEvent | 2023-03-24 15:20:26 | KDT3-Final-6/final-project-FE | https://api.github.com/repos/KDT3-Final-6/final-project-FE | opened | Design: 로그인 & 회원가입 마크업 | Status: Available Status: Review Needed Priority: High | ## ✔️ 체크사항
- [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요.
## 💡 개발 목적
- 로그인 & 회원가입 마크업
## 🌐 세부 내용
- [ ] 회원가입
- [ ] 로그인
## ⚙️ 참조
[피그마] (https://www.figma.com/file/AVUhg8ZkRtoas58FMo95SE/%EA%B3%A0%ED%88%AC%EA%B2%8C%EB%8D%94-6%EC%A1%B0?node-id=1-2&t=PdJGv1AtnUhZyXyp-0)
| 1.0 | Design: 로그인 & 회원가입 마크업 - ## ✔️ 체크사항
- [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요.
## 💡 개발 목적
- 로그인 & 회원가입 마크업
## 🌐 세부 내용
- [ ] 회원가입
- [ ] 로그인
## ⚙️ 참조
[피그마] (https://www.figma.com/file/AVUhg8ZkRtoas58FMo95SE/%EA%B3%A0%ED%88%AC%EA%B2%8C%EB%8D%94-6%EC%A1%B0?node-id=1-2&t=PdJGv1AtnUhZyXyp-0)
| priority | design 로그인 회원가입 마크업 ✔️ 체크사항 제목은 개발 페이지 개발 목적 으로 작성해 주세요 💡 개발 목적 로그인 회원가입 마크업 🌐 세부 내용 회원가입 로그인 ⚙️ 참조 | 1 |
520,057 | 15,077,801,727 | IssuesEvent | 2021-02-05 07:38:57 | wso2/cellery-integration-tests | https://api.github.com/repos/wso2/cellery-integration-tests | closed | Package structure should be revisited and refactored | Priority/High Resolution/Won’t Fix Type/Improvement | **Description:**
The current code has no proper structure and it makes it hard maintain the code. The packaging should be restructured.
| 1.0 | Package structure should be revisited and refactored - **Description:**
The current code has no proper structure and it makes it hard maintain the code. The packaging should be restructured.
| priority | package structure should be revisited and refactored description the current code has no proper structure and it makes it hard maintain the code the packaging should be restructured | 1 |
522,761 | 15,167,424,299 | IssuesEvent | 2021-02-12 17:46:18 | ooni/ooni.org | https://api.github.com/repos/ooni/ooni.org | closed | Update OONI Data Policy | effort/M priority/high website | I'll be updating the OONI Data Policy to reflect the changes made in the latest OONI Probe mobile and desktop app releases.
Specifically, I'll be listing all the analytics tools we're using across platforms, and editing other sections now that we have removed the option to opt-in to IP address collection (and other changes). | 1.0 | Update OONI Data Policy - I'll be updating the OONI Data Policy to reflect the changes made in the latest OONI Probe mobile and desktop app releases.
Specifically, I'll be listing all the analytics tools we're using across platforms, and editing other sections now that we have removed the option to opt-in to IP address collection (and other changes). | priority | update ooni data policy i ll be updating the ooni data policy to reflect the changes made in the latest ooni probe mobile and desktop app releases specifically i ll be listing all the analytics tools we re using across platforms and editing other sections now that we have removed the option to opt in to ip address collection and other changes | 1 |
740,828 | 25,769,876,131 | IssuesEvent | 2022-12-09 06:52:06 | makeniii/NBA-Elo-Rankings | https://api.github.com/repos/makeniii/NBA-Elo-Rankings | closed | 2. Get PR of all NBA teams | Priority: High | Returns a list of all NBA teams sorted by PR.
Acceptance criteria:
- [x] Returns a list of all NBA teams
- [x] For each NBA team
- - [x] Have name of team
- - [x] Current PR
- [x] Sort list by descending PR | 1.0 | 2. Get PR of all NBA teams - Returns a list of all NBA teams sorted by PR.
Acceptance criteria:
- [x] Returns a list of all NBA teams
- [x] For each NBA team
- - [x] Have name of team
- - [x] Current PR
- [x] Sort list by descending PR | priority | get pr of all nba teams returns a list of all nba teams sorted by pr acceptance criteria returns a list of all nba teams for each nba team have name of team current pr sort list by descending pr | 1 |
490,859 | 14,141,031,197 | IssuesEvent | 2020-11-10 12:05:50 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | [ui_next] Surveys cannot be enabled via Survey Toggle for WFJT | component:ui_next priority:high qe:escape qe:regression state:in_progress type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Surveys cannot be enabled via Survey Toggle for WFJT
##### ENVIRONMENT
* AWX version: 13
##### STEPS TO REPRODUCE
1. Create a WFJT
2. Navigate to Survey tab and create a survey
3. Click 'Enable Surveys toggle'
Note: Error message
##### EXPECTED RESULTS
Survey toggled on
##### ACTUAL RESULTS
Survey not toggled on
Error message
##### ADDITIONAL INFORMATION
| 1.0 | [ui_next] Surveys cannot be enabled via Survey Toggle for WFJT - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Surveys cannot be enabled via Survey Toggle for WFJT
##### ENVIRONMENT
* AWX version: 13
##### STEPS TO REPRODUCE
1. Create a WFJT
2. Navigate to Survey tab and create a survey
3. Click 'Enable Surveys toggle'
Note: Error message
##### EXPECTED RESULTS
Survey toggled on
##### ACTUAL RESULTS
Survey not toggled on
Error message
##### ADDITIONAL INFORMATION
| priority | surveys cannot be enabled via survey toggle for wfjt issue type bug report summary surveys cannot be enabled via survey toggle for wfjt environment awx version steps to reproduce create a wfjt navigate to survey tab and create a survey click enable surveys toggle note error message expected results survey toggled on actual results survey not toggled on error message additional information | 1 |
557,265 | 16,505,076,636 | IssuesEvent | 2021-05-25 18:14:54 | airbytehq/airbyte | https://api.github.com/repos/airbytehq/airbyte | opened | regularly run tests on current release | priority/high type/bug | We should be able to know the same day we start encountering issues like https://github.com/airbytehq/airbyte/issues/3588 that stop fresh installs. We shouldn't need to wait until a user reports it. | 1.0 | regularly run tests on current release - We should be able to know the same day we start encountering issues like https://github.com/airbytehq/airbyte/issues/3588 that stop fresh installs. We shouldn't need to wait until a user reports it. | priority | regularly run tests on current release we should be able to know the same day we start encountering issues like that stop fresh installs we shouldn t need to wait until a user reports it | 1 |
178,755 | 6,618,291,944 | IssuesEvent | 2017-09-21 07:30:56 | hobbit-project/platform | https://api.github.com/repos/hobbit-project/platform | opened | Logout not working for anonymous user | component: UI priority: high type: bug | ## Problem
When logged in as anonymous user, the logout button is not working correctly. Instead of being redirected to the login page of Keycloak the user is redirected to the main page of the GUI still logged in as anonymous user. | 1.0 | Logout not working for anonymous user - ## Problem
When logged in as anonymous user, the logout button is not working correctly. Instead of being redirected to the login page of Keycloak the user is redirected to the main page of the GUI still logged in as anonymous user. | priority | logout not working for anonymous user problem when logged in as anonymous user the logout button is not working correctly instead of being redirected to the login page of keycloak the user is redirected to the main page of the gui still logged in as anonymous user | 1 |
714,401 | 24,560,393,126 | IssuesEvent | 2022-10-12 19:41:29 | interaction-lab/MoveToCode | https://api.github.com/repos/interaction-lab/MoveToCode | closed | Behavior Clean Up | high priority | - [x] high five needs sound
- [x] sound for behaviors
- [x] switch mode flashing when should be off
- [x] remove piece back to text
- [x] don't do anything on connect/update action at end of loggable action
- [x] pulse next maze only on high five
- [x] enter goal explode only on correct
- [x] put high five down if press next maze button
- [x] high five no work on end every time
- [x] hold picture above head
- [x] kuri head busted
- [x] ask for help button???
- [x] if kuri colliding eith maze, move away
---
- [x] move then point should be in sequence
- [x] bug on start up run away? -> might just be the ipad...
- [x] look at user on high five
- [x] text too large in front of me
- [x] turn off kuri move when not editor
- [x] too close to stuff
- [x] on reset, reset all anims
- [x] choose side of checker hit
- [x] say something code help
- [x] move away 1 metere from user
- [x] code help should move to an obj
- [x] help button stopped working
- [x] add connectors hit as a human action
- [x] build then code text change on maze correct
- [x] logger runs a ton, not sure why, need to look into this as it is what is stalling out
---
- [x] test out with people
---
### Probably not going to do until playtesting done
- [x] add in state where nothing is happening and deal with high five being put down
- [x] auto clear long text low prio in `KuriTextManager.cs`
| 1.0 | Behavior Clean Up - - [x] high five needs sound
- [x] sound for behaviors
- [x] switch mode flashing when should be off
- [x] remove piece back to text
- [x] don't do anything on connect/update action at end of loggable action
- [x] pulse next maze only on high five
- [x] enter goal explode only on correct
- [x] put high five down if press next maze button
- [x] high five no work on end every time
- [x] hold picture above head
- [x] kuri head busted
- [x] ask for help button???
- [x] if kuri colliding eith maze, move away
---
- [x] move then point should be in sequence
- [x] bug on start up run away? -> might just be the ipad...
- [x] look at user on high five
- [x] text too large in front of me
- [x] turn off kuri move when not editor
- [x] too close to stuff
- [x] on reset, reset all anims
- [x] choose side of checker hit
- [x] say something code help
- [x] move away 1 metere from user
- [x] code help should move to an obj
- [x] help button stopped working
- [x] add connectors hit as a human action
- [x] build then code text change on maze correct
- [x] logger runs a ton, not sure why, need to look into this as it is what is stalling out
---
- [x] test out with people
---
### Probably not going to do until playtesting done
- [x] add in state where nothing is happening and deal with high five being put down
- [x] auto clear long text low prio in `KuriTextManager.cs`
| priority | behavior clean up high five needs sound sound for behaviors switch mode flashing when should be off remove piece back to text don t do anything on connect update action at end of loggable action pulse next maze only on high five enter goal explode only on correct put high five down if press next maze button high five no work on end every time hold picture above head kuri head busted ask for help button if kuri colliding eith maze move away move then point should be in sequence bug on start up run away might just be the ipad look at user on high five text too large in front of me turn off kuri move when not editor too close to stuff on reset reset all anims choose side of checker hit say something code help move away metere from user code help should move to an obj help button stopped working add connectors hit as a human action build then code text change on maze correct logger runs a ton not sure why need to look into this as it is what is stalling out test out with people probably not going to do until playtesting done add in state where nothing is happening and deal with high five being put down auto clear long text low prio in kuritextmanager cs | 1 |
782,447 | 27,496,913,941 | IssuesEvent | 2023-03-05 08:38:05 | ThomUK/SPCreporter | https://api.github.com/repos/ThomUK/SPCreporter | opened | Implement a new unit type: 'rate per 1000' | enhancement high priority | Some proportion measures need reporting as a rate per 1000, not as a percentage.
Implement rounding for the header row, for both the target, and the last data point | 1.0 | Implement a new unit type: 'rate per 1000' - Some proportion measures need reporting as a rate per 1000, not as a percentage.
Implement rounding for the header row, for both the target, and the last data point | priority | implement a new unit type rate per some proportion measures need reporting as a rate per not as a percentage implement rounding for the header row for both the target and the last data point | 1 |
133,984 | 5,218,019,273 | IssuesEvent | 2017-01-26 15:29:41 | easydigitaldownloads/easy-digital-downloads | https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads | closed | Turn edd_get_payment into a caching layer for EDD_Payment, and allow false on no-payment. | Bug Payments Priority: High | When passing a `$payment_id` that isn't a valid payment, `EDD_Payment::setup_payment()` returns false before setting up the payment.
`edd_update_payment_status()` assumes that instantiating `EDD_Payment` returns a valid `EDD_Payment` class, and continues with that assumption, calling `EDD_Payment::save()` no matter the return value.
Related to https://trello.com/c/j7LgHA80/436-gravity-forms-checkout-for-easy-digital-downloads#comment-575edeabeb0ad751c227c4e9
| 1.0 | Turn edd_get_payment into a caching layer for EDD_Payment, and allow false on no-payment. - When passing a `$payment_id` that isn't a valid payment, `EDD_Payment::setup_payment()` returns false before setting up the payment.
`edd_update_payment_status()` assumes that instantiating `EDD_Payment` returns a valid `EDD_Payment` class, and continues with that assumption, calling `EDD_Payment::save()` no matter the return value.
Related to https://trello.com/c/j7LgHA80/436-gravity-forms-checkout-for-easy-digital-downloads#comment-575edeabeb0ad751c227c4e9
| priority | turn edd get payment into a caching layer for edd payment and allow false on no payment when passing a payment id that isn t a valid payment edd payment setup payment returns false before setting up the payment edd update payment status assumes that instantiating edd payment returns a valid edd payment class and continues with that assumption calling edd payment save no matter the return value related to | 1 |
364,377 | 10,763,338,767 | IssuesEvent | 2019-11-01 03:32:15 | ShabadOS/desktop | https://api.github.com/repos/ShabadOS/desktop | closed | feat(frontend/settings): add highlight current line option | Priority: High Status: Confirmed Type: Feature | With the introduction of multiple previous+next lines, it can be confusing to determine what the current line activated is. I suggest adding a "{highlight, accent, contrasting} current line" option with some CSS it would be easier to know what the main line is. I can provide the CSS on a theme-by-theme basis if the react component is set up properly. | 1.0 | feat(frontend/settings): add highlight current line option - With the introduction of multiple previous+next lines, it can be confusing to determine what the current line activated is. I suggest adding a "{highlight, accent, contrasting} current line" option with some CSS it would be easier to know what the main line is. I can provide the CSS on a theme-by-theme basis if the react component is set up properly. | priority | feat frontend settings add highlight current line option with the introduction of multiple previous next lines it can be confusing to determine what the current line activated is i suggest adding a highlight accent contrasting current line option with some css it would be easier to know what the main line is i can provide the css on a theme by theme basis if the react component is set up properly | 1 |
673,772 | 23,030,396,796 | IssuesEvent | 2022-07-22 13:24:59 | canonical/prometheus-k8s-operator | https://api.github.com/repos/canonical/prometheus-k8s-operator | closed | MetricsEndpointProvider units are not always reachable via fqdn | Priority: High Type: Bug Area: prometheus_scrape | ### Bug Description
#293 replaced `network.bind_address()` with `socket.getfqdn()`. This works for in-model relations but breaks for cross-cluster relations.
cc: @mateoflorido @stonepreston
### To Reproduce
Form a cross-cluster prometheus_scrape relation b/w a machine charm and the prometheus-k8s charm.
### Environment
> we have been trying to integrate the prometheus-k8s charm with kube-ovn charm. But it seems that the library gathers the instance-id of our charm instead of the IP address of the unit, therefore metrics are not being collected in the process.
>
> they are in different models. We have a CK cluster that has the kube-ovn charm as it's CNI charm.
>
> Kube-OVN is a subordinate machine charm.
> -- @mateoflorido
### Relevant log output
```shell
~$ curl 10.152.183.72:9090/api/v1/targets
{
"status": "success",
"data": {
"activeTargets": [
{
"discoveredLabels": {
"__address__": "juju-865c1a-5:10665",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/4",
},
"labels": {
"instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/4",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/4",
},
"scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"scrapeUrl": "http://juju-865c1a-5:10665/metrics",
"globalUrl": "http://juju-865c1a-5:10665/metrics",
"lastError": 'Get "http://juju-865c1a-5:10665/metrics": dial tcp: lookup juju-865c1a-5 on 10.152.183.213:53: no such host',
"lastScrape": "2022-07-15T15:27:59.615508445Z",
"lastScrapeDuration": 0.005049553,
"health": "down",
},
{
"discoveredLabels": {
"__address__": "juju-865c1a-8:10665",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/0",
},
"labels": {
"instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/0",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/0",
},
"scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"scrapeUrl": "http://juju-865c1a-8:10665/metrics",
"globalUrl": "http://juju-865c1a-8:10665/metrics",
"lastError": 'Get "http://juju-865c1a-8:10665/metrics": dial tcp: lookup juju-865c1a-8 on 10.152.183.213:53: no such host',
"lastScrape": "2022-07-15T15:28:30.262574207Z",
"lastScrapeDuration": 0.00596523,
"health": "down",
},
{
"discoveredLabels": {
"__address__": "localhost:9090",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "prometheus",
},
"labels": {"instance": "localhost:9090", "job": "prometheus"},
"scrapePool": "prometheus",
"scrapeUrl": "http://localhost:9090/metrics",
"globalUrl": "http://192.168.0.17:9090/metrics",
"lastError": "",
"lastScrape": "2022-07-15T15:28:46.369824673Z",
"lastScrapeDuration": 0.006802522,
"health": "up",
},
],
"droppedTargets": [],
},
}
```
### Additional context
_No response_ | 1.0 | MetricsEndpointProvider units are not always reachable via fqdn - ### Bug Description
#293 replaced `network.bind_address()` with `socket.getfqdn()`. This works for in-model relations but breaks for cross-cluster relations.
cc: @mateoflorido @stonepreston
### To Reproduce
Form a cross-cluster prometheus_scrape relation b/w a machine charm and the prometheus-k8s charm.
### Environment
> we have been trying to integrate the prometheus-k8s charm with kube-ovn charm. But it seems that the library gathers the instance-id of our charm instead of the IP address of the unit, therefore metrics are not being collected in the process.
>
> they are in different models. We have a CK cluster that has the kube-ovn charm as it's CNI charm.
>
> Kube-OVN is a subordinate machine charm.
> -- @mateoflorido
### Relevant log output
```shell
~$ curl 10.152.183.72:9090/api/v1/targets
{
"status": "success",
"data": {
"activeTargets": [
{
"discoveredLabels": {
"__address__": "juju-865c1a-5:10665",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/4",
},
"labels": {
"instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/4",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/4",
},
"scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"scrapeUrl": "http://juju-865c1a-5:10665/metrics",
"globalUrl": "http://juju-865c1a-5:10665/metrics",
"lastError": 'Get "http://juju-865c1a-5:10665/metrics": dial tcp: lookup juju-865c1a-5 on 10.152.183.213:53: no such host',
"lastScrape": "2022-07-15T15:27:59.615508445Z",
"lastScrapeDuration": 0.005049553,
"health": "down",
},
{
"discoveredLabels": {
"__address__": "juju-865c1a-8:10665",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/0",
},
"labels": {
"instance": "test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn/0",
"job": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"juju_application": "kube-ovn",
"juju_charm": "kube-ovn",
"juju_model": "test-kovn-prom",
"juju_model_uuid": "e7559de7-acde-415f-84fa-eeff47865c1a",
"juju_unit": "kube-ovn/0",
},
"scrapePool": "juju_test-kovn-prom_e7559de7-acde-415f-84fa-eeff47865c1a_kube-ovn_kube-ovn_prometheus_scrape",
"scrapeUrl": "http://juju-865c1a-8:10665/metrics",
"globalUrl": "http://juju-865c1a-8:10665/metrics",
"lastError": 'Get "http://juju-865c1a-8:10665/metrics": dial tcp: lookup juju-865c1a-8 on 10.152.183.213:53: no such host',
"lastScrape": "2022-07-15T15:28:30.262574207Z",
"lastScrapeDuration": 0.00596523,
"health": "down",
},
{
"discoveredLabels": {
"__address__": "localhost:9090",
"__metrics_path__": "/metrics",
"__scheme__": "http",
"job": "prometheus",
},
"labels": {"instance": "localhost:9090", "job": "prometheus"},
"scrapePool": "prometheus",
"scrapeUrl": "http://localhost:9090/metrics",
"globalUrl": "http://192.168.0.17:9090/metrics",
"lastError": "",
"lastScrape": "2022-07-15T15:28:46.369824673Z",
"lastScrapeDuration": 0.006802522,
"health": "up",
},
],
"droppedTargets": [],
},
}
```
### Additional context
_No response_ | priority | metricsendpointprovider units are not always reachable via fqdn bug description replaced network bind address with socket getfqdn this works for in model relations but breaks for cross cluster relations cc mateoflorido stonepreston to reproduce form a cross cluster prometheus scrape relation b w a machine charm and the prometheus charm environment we have been trying to integrate the prometheus charm with kube ovn charm but it seems that the library gathers the instance id of our charm instead of the ip address of the unit therefore metrics are not being collected in the process they are in different models we have a ck cluster that has the kube ovn charm as it s cni charm kube ovn is a subordinate machine charm mateoflorido relevant log output shell curl api targets status success data activetargets discoveredlabels address juju metrics path metrics scheme http job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn labels instance test kovn prom acde kube ovn kube ovn job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn scrapepool juju test kovn prom acde kube ovn kube ovn prometheus scrape scrapeurl globalurl lasterror get dial tcp lookup juju on no such host lastscrape lastscrapeduration health down discoveredlabels address juju metrics path metrics scheme http job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn labels instance test kovn prom acde kube ovn kube ovn job juju test kovn prom acde kube ovn kube ovn prometheus scrape juju application kube ovn juju charm kube ovn juju model test kovn prom juju model uuid acde juju unit kube ovn scrapepool juju test kovn prom acde kube ovn kube ovn prometheus scrape scrapeurl globalurl lasterror get dial tcp lookup juju on no such host lastscrape lastscrapeduration health down discoveredlabels address localhost metrics path metrics scheme http job prometheus labels instance localhost job prometheus scrapepool prometheus scrapeurl globalurl lasterror lastscrape lastscrapeduration health up droppedtargets additional context no response | 1 |
119,700 | 4,774,618,360 | IssuesEvent | 2016-10-27 07:31:43 | ponylang/ponyc | https://api.github.com/repos/ponylang/ponyc | closed | ponyc built on OS X with llvm-config 3.8 doesn't compile hello world | bug: 1 - needs investigation priority: 3 - high | I pulled master (as of this writing: e4427917a3cccc04dd3633777a31e17b5133a743) and could not build hello world on OS X (El Capitan 10.11.4) when `ponyc` is compiled with LLVM 3.8.
I have both LLVM 3.6.2 and LLVM 3.8 installed with homebrew. I'm not sure if anything other than the location of `llvm-config` matters here.
Compiling hello world works after compiling pony with LLVM 3.6.2 like so:
```
LLVM_CONFIG=/usr/local/opt/llvm/bin/llvm-config make config=release
```
But it does not work after compiling pony with LLVM 3.8 like so:
```
LLVM_CONFIG=/usr/local/bin/llvm-config-3.8 make config=release
```
With `ponyc` compiled with LLVM 3.8
```
ponyc:$ cd examples/helloworld/
helloworld:$ ../../build/release/ponyc
Building builtin -> /Users/jeremy/src/ponyc/packages/builtin
Building . -> /Users/jeremy/src/ponyc/examples/helloworld
Generating
Reachability
Selector painting
Data prototypes
Data types
Function prototypes
Descriptors
Functions
Optimising
Writing ./helloworld.o
Cannot select: t19: ch = store<ST1[%30]> t16, t7, t18, undef:i64
t7: i1,ch = load<LD1[%23]> t0, t4, undef:i64
t4: i64 = add t2, Constant:i64<16>
t2: i64,ch = CopyFromReg t0, Register:i64 %vreg4
t1: i64 = Register %vreg4
t3: i64 = Constant<16>
t6: i64 = undef
t18: i64 = add t9, Constant:i64<264>
t9: i64,ch = CopyFromReg t0, Register:i64 %vreg3
t8: i64 = Register %vreg3
t17: i64 = Constant<264>
t6: i64 = undef
In function: Stdin_$dispatch
``` | 1.0 | ponyc built on OS X with llvm-config 3.8 doesn't compile hello world - I pulled master (as of this writing: e4427917a3cccc04dd3633777a31e17b5133a743) and could not build hello world on OS X (El Capitan 10.11.4) when `ponyc` is compiled with LLVM 3.8.
I have both LLVM 3.6.2 and LLVM 3.8 installed with homebrew. I'm not sure if anything other than the location of `llvm-config` matters here.
Compiling hello world works after compiling pony with LLVM 3.6.2 like so:
```
LLVM_CONFIG=/usr/local/opt/llvm/bin/llvm-config make config=release
```
But it does not work after compiling pony with LLVM 3.8 like so:
```
LLVM_CONFIG=/usr/local/bin/llvm-config-3.8 make config=release
```
With `ponyc` compiled with LLVM 3.8
```
ponyc:$ cd examples/helloworld/
helloworld:$ ../../build/release/ponyc
Building builtin -> /Users/jeremy/src/ponyc/packages/builtin
Building . -> /Users/jeremy/src/ponyc/examples/helloworld
Generating
Reachability
Selector painting
Data prototypes
Data types
Function prototypes
Descriptors
Functions
Optimising
Writing ./helloworld.o
Cannot select: t19: ch = store<ST1[%30]> t16, t7, t18, undef:i64
t7: i1,ch = load<LD1[%23]> t0, t4, undef:i64
t4: i64 = add t2, Constant:i64<16>
t2: i64,ch = CopyFromReg t0, Register:i64 %vreg4
t1: i64 = Register %vreg4
t3: i64 = Constant<16>
t6: i64 = undef
t18: i64 = add t9, Constant:i64<264>
t9: i64,ch = CopyFromReg t0, Register:i64 %vreg3
t8: i64 = Register %vreg3
t17: i64 = Constant<264>
t6: i64 = undef
In function: Stdin_$dispatch
``` | priority | ponyc built on os x with llvm config doesn t compile hello world i pulled master as of this writing and could not build hello world on os x el capitan when ponyc is compiled with llvm i have both llvm and llvm installed with homebrew i m not sure if anything other than the location of llvm config matters here compiling hello world works after compiling pony with llvm like so llvm config usr local opt llvm bin llvm config make config release but it does not work after compiling pony with llvm like so llvm config usr local bin llvm config make config release with ponyc compiled with llvm ponyc cd examples helloworld helloworld build release ponyc building builtin users jeremy src ponyc packages builtin building users jeremy src ponyc examples helloworld generating reachability selector painting data prototypes data types function prototypes descriptors functions optimising writing helloworld o cannot select ch store undef ch load undef add constant ch copyfromreg register register constant undef add constant ch copyfromreg register register constant undef in function stdin dispatch | 1 |
728,242 | 25,072,617,266 | IssuesEvent | 2022-11-07 13:20:08 | lazarskiopencourses/lazarskiopencourses.github.io | https://api.github.com/repos/lazarskiopencourses/lazarskiopencourses.github.io | opened | Programming in Python - Advanced Function - Recursion - Fibonacci | bug high priority | There is a mistake in the example code on the website --> `fib(6)` should return `(8, 25)` | 1.0 | Programming in Python - Advanced Function - Recursion - Fibonacci - There is a mistake in the example code on the website --> `fib(6)` should return `(8, 25)` | priority | programming in python advanced function recursion fibonacci there is a mistake in the example code on the website fib should return | 1 |
761,975 | 26,705,210,507 | IssuesEvent | 2023-01-27 17:31:29 | evo-lua/evo-runtime | https://api.github.com/repos/evo-lua/evo-runtime | opened | Re-add the event system library so that events can be monitored (for testing purposes) | Priority: High Complexity: Low Scope: Runtime Status: Accepted Type: New Feature | This is the global event registry from evo-luvi (see RFC), but it should be able to buffer events and payloads for testing purposes.
Using event emitters makes the code less decoupled and it's a bit of a pain to test whether a server has received some data, for example, which was a problem I encountered when writing tests for the TCP server before. I basically had to write a lot of boilerplate code, when I think it might be easier to just enable the event buffer, do the thing, and then fetch the events and compare to a list (TBD). | 1.0 | Re-add the event system library so that events can be monitored (for testing purposes) - This is the global event registry from evo-luvi (see RFC), but it should be able to buffer events and payloads for testing purposes.
Using event emitters makes the code less decoupled and it's a bit of a pain to test whether a server has received some data, for example, which was a problem I encountered when writing tests for the TCP server before. I basically had to write a lot of boilerplate code, when I think it might be easier to just enable the event buffer, do the thing, and then fetch the events and compare to a list (TBD). | priority | re add the event system library so that events can be monitored for testing purposes this is the global event registry from evo luvi see rfc but it should be able to buffer events and payloads for testing purposes using event emitters makes the code less decoupled and it s a bit of a pain to test whether a server has received some data for example which was a problem i encountered when writing tests for the tcp server before i basically had to write a lot of boilerplate code when i think it might be easier to just enable the event buffer do the thing and then fetch the events and compare to a list tbd | 1 |
283,987 | 8,729,004,023 | IssuesEvent | 2018-12-10 19:00:09 | conveyal/datatools-ui | https://api.github.com/repos/conveyal/datatools-ui | closed | OTP Deployment: Bounds should be checked for validity before completing deployment | bug high-priority imported | <a href="https://github.com/landonreed"><img src="https://avatars2.githubusercontent.com/u/2370911?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [landonreed](https://github.com/landonreed)**
_Friday Jun 01, 2018 at 19:08 GMT_
_Originally opened as https://github.com/catalogueglobal/datatools-ui/issues/189_
----
A recent deployment job succeeded even though the osm extract was for the bounds `vex/NaN,NaN,NaN,NaN.pbf`. A couple of fixes might be in order:
- The server should check that the bounds are valid and fail if not, so that an incomplete set of data is not sent for graph building.
- The UI should warn the user that the project bounds are invalid and which feed versions might be problematic.
| 1.0 | OTP Deployment: Bounds should be checked for validity before completing deployment - <a href="https://github.com/landonreed"><img src="https://avatars2.githubusercontent.com/u/2370911?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [landonreed](https://github.com/landonreed)**
_Friday Jun 01, 2018 at 19:08 GMT_
_Originally opened as https://github.com/catalogueglobal/datatools-ui/issues/189_
----
A recent deployment job succeeded even though the osm extract was for the bounds `vex/NaN,NaN,NaN,NaN.pbf`. A couple of fixes might be in order:
- The server should check that the bounds are valid and fail if not, so that an incomplete set of data is not sent for graph building.
- The UI should warn the user that the project bounds are invalid and which feed versions might be problematic.
| priority | otp deployment bounds should be checked for validity before completing deployment issue by friday jun at gmt originally opened as a recent deployment job succeeded even though the osm extract was for the bounds vex nan nan nan nan pbf a couple of fixes might be in order the server should check that the bounds are valid and fail if not so that an incomplete set of data is not sent for graph building the ui should warn the user that the project bounds are invalid and which feed versions might be problematic | 1 |
396,504 | 11,709,729,755 | IssuesEvent | 2020-03-08 20:29:56 | open-gunz/source | https://api.github.com/repos/open-gunz/source | opened | Server-based netcode and AntiLead | High Priority | Finish the server-based netcode that crawly has made to have real-time discerning between sword hits/rockets/grenades/hitscan weapons/etc. NAT - server sided hit detection.
Current Anti-lead issues: Packets sometimes get lost/malformed that can happen once every few 100 packets with UDP. This leads to showing HIT in P2P anti-lead but still does not reduce target's HP/AP. Sometimes, there is no HIT but the target's HP/AP is reduced. | 1.0 | Server-based netcode and AntiLead - Finish the server-based netcode that crawly has made to have real-time discerning between sword hits/rockets/grenades/hitscan weapons/etc. NAT - server sided hit detection.
Current Anti-lead issues: Packets sometimes get lost/malformed that can happen once every few 100 packets with UDP. This leads to showing HIT in P2P anti-lead but still does not reduce target's HP/AP. Sometimes, there is no HIT but the target's HP/AP is reduced. | priority | server based netcode and antilead finish the server based netcode that crawly has made to have real time discerning between sword hits rockets grenades hitscan weapons etc nat server sided hit detection current anti lead issues packets sometimes get lost malformed that can happen once every few packets with udp this leads to showing hit in anti lead but still does not reduce target s hp ap sometimes there is no hit but the target s hp ap is reduced | 1 |
517,138 | 14,995,403,885 | IssuesEvent | 2021-01-29 14:16:39 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Photo more options hover, photo popup shows previous image for a while | bug priority: high | **Describe the bug**
1. Photos (...) ellipsis dropdown hover is not proper
2. Photos Popup behaving weird by showing last open image and then new select image.
3. View Group Photos in Popup as Organizer > Fix order here to show “Report” just before “delete” button: https://prnt.sc/xpffcs
**Video**
https://www.loom.com/share/8a9e51400ce2466387ce520f31be768d | 1.0 | Photo more options hover, photo popup shows previous image for a while - **Describe the bug**
1. Photos (...) ellipsis dropdown hover is not proper
2. Photos Popup behaving weird by showing last open image and then new select image.
3. View Group Photos in Popup as Organizer > Fix order here to show “Report” just before “delete” button: https://prnt.sc/xpffcs
**Video**
https://www.loom.com/share/8a9e51400ce2466387ce520f31be768d | priority | photo more options hover photo popup shows previous image for a while describe the bug photos ellipsis dropdown hover is not proper photos popup behaving weird by showing last open image and then new select image view group photos in popup as organizer fix order here to show “report” just before “delete” button video | 1 |
89,289 | 3,792,076,257 | IssuesEvent | 2016-03-22 07:49:16 | cs2103jan2016-W10-1J/main | https://api.github.com/repos/cs2103jan2016-W10-1J/main | closed | Logic support according to Architectural change adopting inheritance and polymorphism | priority.high | - [x] Code for Update Class as part of the Logic
- [x] Code for the Add Class as part of the Logic | 1.0 | Logic support according to Architectural change adopting inheritance and polymorphism - - [x] Code for Update Class as part of the Logic
- [x] Code for the Add Class as part of the Logic | priority | logic support according to architectural change adopting inheritance and polymorphism code for update class as part of the logic code for the add class as part of the logic | 1 |
172,258 | 6,500,824,159 | IssuesEvent | 2017-08-23 07:02:01 | ballerinalang/composer | https://api.github.com/repos/ballerinalang/composer | opened | [xmlLiterals] Giving syntax errors as \ will be removed | Bug Priority:High Severity:Major | Pack - 23/08/2017
This given as error as the slash (\) will be removed when you switch to the sources view adding this from the design view
`xml x5 = xml `aaa{{v1}}b\{{bb{{v2}}c\}}cc{d{}e}{f{`;` | 1.0 | [xmlLiterals] Giving syntax errors as \ will be removed - Pack - 23/08/2017
This given as error as the slash (\) will be removed when you switch to the sources view adding this from the design view
`xml x5 = xml `aaa{{v1}}b\{{bb{{v2}}c\}}cc{d{}e}{f{`;` | priority | giving syntax errors as will be removed pack this given as error as the slash will be removed when you switch to the sources view adding this from the design view xml xml aaa b bb c cc d e f | 1 |
439,159 | 12,678,386,926 | IssuesEvent | 2020-06-19 09:40:50 | localstack/localstack | https://api.github.com/repos/localstack/localstack | closed | Unable to determine forwarding port for API "monitoring" | bug priority-high | <!-- Love localstack? Please consider supporting our collective:
👉 https://opencollective.com/localstack/donate -->
# Type of request: This is a ...
[x ] bug report
[ ] feature request
# Detailed description
Not able to use the edge service to write metric data to cloudwatch via the javascript aws-sdk.
However, It **DOES** work when using the deprecated cloudwatch port.
## Expected behavior
Metric data to be visible when calling `awslocal cloudwatch list-metrics`
## Actual behavior
404 from edge service. No metric data is persisted, localstack instance logs the following info:
```
INFO:localstack.services.edge: Unable to determine forwarding port for API "monitoring" - please make sure this API is enabled via the SERVICES configuration
```
# Steps to reproduce
Run JS code in "Client code" section.
See 404 Error
## Command used to start LocalStack
`TMPDIR=/tmp/localstack docker-compose up`
## Client code (AWS SDK code snippet, or sequence of "awslocal" commands)
```js
const AWS = require('aws-sdk')
const cw = new AWS.CloudWatch({
endpoint: 'http://localhost:4566',
})
const data = {
Namespace: 'ns',
MetricData: [
{
MetricName: 'metric-name',
Timestamp: new Date(),
Value: 12345,
Unit: 'Milliseconds',
Dimensions: [
{ Name: 'Environment', Value: 'local' },
]
}
]
}
cw.putMetricData(data, (err, data) => {
if (err) {
console.log('Error', err)
} else {
console.log('Success', JSON.stringify(data))
}
})
```
| 1.0 | Unable to determine forwarding port for API "monitoring" - <!-- Love localstack? Please consider supporting our collective:
👉 https://opencollective.com/localstack/donate -->
# Type of request: This is a ...
[x ] bug report
[ ] feature request
# Detailed description
Not able to use the edge service to write metric data to cloudwatch via the javascript aws-sdk.
However, It **DOES** work when using the deprecated cloudwatch port.
## Expected behavior
Metric data to be visible when calling `awslocal cloudwatch list-metrics`
## Actual behavior
404 from edge service. No metric data is persisted, localstack instance logs the following info:
```
INFO:localstack.services.edge: Unable to determine forwarding port for API "monitoring" - please make sure this API is enabled via the SERVICES configuration
```
# Steps to reproduce
Run JS code in "Client code" section.
See 404 Error
## Command used to start LocalStack
`TMPDIR=/tmp/localstack docker-compose up`
## Client code (AWS SDK code snippet, or sequence of "awslocal" commands)
```js
const AWS = require('aws-sdk')
const cw = new AWS.CloudWatch({
endpoint: 'http://localhost:4566',
})
const data = {
Namespace: 'ns',
MetricData: [
{
MetricName: 'metric-name',
Timestamp: new Date(),
Value: 12345,
Unit: 'Milliseconds',
Dimensions: [
{ Name: 'Environment', Value: 'local' },
]
}
]
}
cw.putMetricData(data, (err, data) => {
if (err) {
console.log('Error', err)
} else {
console.log('Success', JSON.stringify(data))
}
})
```
| priority | unable to determine forwarding port for api monitoring love localstack please consider supporting our collective 👉 type of request this is a bug report feature request detailed description not able to use the edge service to write metric data to cloudwatch via the javascript aws sdk however it does work when using the deprecated cloudwatch port expected behavior metric data to be visible when calling awslocal cloudwatch list metrics actual behavior from edge service no metric data is persisted localstack instance logs the following info info localstack services edge unable to determine forwarding port for api monitoring please make sure this api is enabled via the services configuration steps to reproduce run js code in client code section see error command used to start localstack tmpdir tmp localstack docker compose up client code aws sdk code snippet or sequence of awslocal commands js const aws require aws sdk const cw new aws cloudwatch endpoint const data namespace ns metricdata metricname metric name timestamp new date value unit milliseconds dimensions name environment value local cw putmetricdata data err data if err console log error err else console log success json stringify data | 1 |
439,359 | 12,681,723,483 | IssuesEvent | 2020-06-19 15:55:20 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | opened | Task stats in User DTO | Component: Backend Difficulty: 3 Priority: High Type: Bug | `/users/queries/username` endpoint is missing details on task stats - mapped/validated/invalidated | 1.0 | Task stats in User DTO - `/users/queries/username` endpoint is missing details on task stats - mapped/validated/invalidated | priority | task stats in user dto users queries username endpoint is missing details on task stats mapped validated invalidated | 1 |
281,402 | 8,694,755,567 | IssuesEvent | 2018-12-04 13:33:13 | cilium/cilium | https://api.github.com/repos/cilium/cilium | closed | toFQDNs with no IP data do not block traffic | area/daemon area/policy kind/bug priority/high | When a toFQDN rule has L4/L7 sections, it will incorrectly interpret the L3 as a wildcard allow-all. This means it isn't really enforcing what one expects. When IPs are inserted into toCIDRSet the enforcement is correct.
This policy works incorrectly with our minikube GSG:
```
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: rebel-escape
namespace: default
spec:
endpointSelector:
matchLabels:
any:org: alliance
egress:
- toEndpoints:
- matchLabels:
"k8s:io.kubernetes.pod.namespace": kube-system
"k8s:k8s-app": kube-dns
- toFQDNs:
- matchName: "notasubdomain.cilium.io"
toPorts:
- ports:
- port: "80"
```
Both of these work when they should not (the port is enforced, as would a L7 HTTP rules)
`kubectl exec xwing -- curl -s cilium.io`
`kubectl exec xwing -- curl -s google.com` | 1.0 | toFQDNs with no IP data do not block traffic - When a toFQDN rule has L4/L7 sections, it will incorrectly interpret the L3 as a wildcard allow-all. This means it isn't really enforcing what one expects. When IPs are inserted into toCIDRSet the enforcement is correct.
This policy works incorrectly with our minikube GSG:
```
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: rebel-escape
namespace: default
spec:
endpointSelector:
matchLabels:
any:org: alliance
egress:
- toEndpoints:
- matchLabels:
"k8s:io.kubernetes.pod.namespace": kube-system
"k8s:k8s-app": kube-dns
- toFQDNs:
- matchName: "notasubdomain.cilium.io"
toPorts:
- ports:
- port: "80"
```
Both of these work when they should not (the port is enforced, as would a L7 HTTP rules)
`kubectl exec xwing -- curl -s cilium.io`
`kubectl exec xwing -- curl -s google.com` | priority | tofqdns with no ip data do not block traffic when a tofqdn rule has sections it will incorrectly interpret the as a wildcard allow all this means it isn t really enforcing what one expects when ips are inserted into tocidrset the enforcement is correct this policy works incorrectly with our minikube gsg apiversion cilium io kind ciliumnetworkpolicy metadata name rebel escape namespace default spec endpointselector matchlabels any org alliance egress toendpoints matchlabels io kubernetes pod namespace kube system app kube dns tofqdns matchname notasubdomain cilium io toports ports port both of these work when they should not the port is enforced as would a http rules kubectl exec xwing curl s cilium io kubectl exec xwing curl s google com | 1 |
805,460 | 29,520,516,223 | IssuesEvent | 2023-06-05 01:00:48 | steedos/steedos-platform | https://api.github.com/repos/steedos/steedos-platform | closed | [Bug]: 审批,待审核/监控箱等列表,搜素,流程字段统一为新建流程的式样 | bug done priority: High | ### Description
目前是普通lookup类型字段的式样

应统一为 新建流程时的选项式样

BTW,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些
### Steps To Reproduce 重现步骤
审批,待审核/监控箱等列表,搜素,流程
### Version 版本
2.5.0-beta.39 | 1.0 | [Bug]: 审批,待审核/监控箱等列表,搜素,流程字段统一为新建流程的式样 - ### Description
目前是普通lookup类型字段的式样

应统一为 新建流程时的选项式样

BTW,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些
### Steps To Reproduce 重现步骤
审批,待审核/监控箱等列表,搜素,流程
### Version 版本
2.5.0-beta.39 | priority | 审批,待审核 监控箱等列表,搜素,流程字段统一为新建流程的式样 description 目前是普通lookup类型字段的式样 应统一为 新建流程时的选项式样 btw,新建列出的仅本人有权限提交的,而待办、监控里其应不仅这些 steps to reproduce 重现步骤 审批,待审核 监控箱等列表,搜素,流程 version 版本 beta | 1 |
288,869 | 8,852,522,493 | IssuesEvent | 2019-01-08 18:36:14 | visit-dav/issues-test | https://api.github.com/repos/visit-dav/issues-test | closed | VisIt crashes on startup on Windows Vista | bug crash likelihood high priority reviewed severity high wrong results | When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 192
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt crashes on startup on Windows Vista
Assigned to: Kathleen Biagas
Category: -
Target version: 2.1
Author: Kathleen Biagas
Start: 06/30/2010
Due date:
% Done: 0%
Estimated time:
Created: 06/30/2010 08:19 pm
Updated: 08/27/2010 05:34 pm
Likelihood: 4 - Common
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.0
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2.
Comments:
Assignment from LLNL VisIt 2.1 Release Meeting I downloaded Microsoft's 'Application Compatibility Toolkit' in order to track down why Vista is flagging VisIt as needing to be run in compatibility mode (especially an NT4 compat mode!)First few passes with the tool appear to indicate that VisIt is attempting to WRITE to HKLM registry files, which should notbe happening. That's the only compatibility issues that cropped up. The registry write does not appear in VisIt's source code.I ran VisIt through a a tool that generates callstack information to discover the source of the Registry write operations, andit appears to be happening down in GL calls. (wglSwapMultipleBuffers). If this is truly the case, I'm not sure what we can doto mitigate this issue on Vista. Binaries built with Visual Studio 9.0 (2008) do not have the same issue.Also, binaries built on Vista using Visual Studio 8 do not have this issue.Starting with VisIt 2.1, we will distribute binaries built with Visual Studio 9.
| 1.0 | VisIt crashes on startup on Windows Vista - When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 192
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt crashes on startup on Windows Vista
Assigned to: Kathleen Biagas
Category: -
Target version: 2.1
Author: Kathleen Biagas
Start: 06/30/2010
Due date:
% Done: 0%
Estimated time:
Created: 06/30/2010 08:19 pm
Updated: 08/27/2010 05:34 pm
Likelihood: 4 - Common
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.0
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
When starting VisIt on Windows Vista, the mdserver crashes on startup. A few users have reported this workaround: to run VisIt with compatibility mode set to NT 4 (service pack 5).I verified this work-around on my version of Vista, which is Vista Business 64 bit, SP2.
Comments:
Assignment from LLNL VisIt 2.1 Release Meeting I downloaded Microsoft's 'Application Compatibility Toolkit' in order to track down why Vista is flagging VisIt as needing to be run in compatibility mode (especially an NT4 compat mode!)First few passes with the tool appear to indicate that VisIt is attempting to WRITE to HKLM registry files, which should notbe happening. That's the only compatibility issues that cropped up. The registry write does not appear in VisIt's source code.I ran VisIt through a a tool that generates callstack information to discover the source of the Registry write operations, andit appears to be happening down in GL calls. (wglSwapMultipleBuffers). If this is truly the case, I'm not sure what we can doto mitigate this issue on Vista. Binaries built with Visual Studio 9.0 (2008) do not have the same issue.Also, binaries built on Vista using Visual Studio 8 do not have this issue.Starting with VisIt 2.1, we will distribute binaries built with Visual Studio 9.
| priority | visit crashes on startup on windows vista when starting visit on windows vista the mdserver crashes on startup a few users have reported this workaround to run visit with compatibility mode set to nt service pack i verified this work around on my version of vista which is vista business bit redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit crashes on startup on windows vista assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood common severity crash wrong results found in version impact expected use os windows support group any description when starting visit on windows vista the mdserver crashes on startup a few users have reported this workaround to run visit with compatibility mode set to nt service pack i verified this work around on my version of vista which is vista business bit comments assignment from llnl visit release meeting i downloaded microsoft s application compatibility toolkit in order to track down why vista is flagging visit as needing to be run in compatibility mode especially an compat mode first few passes with the tool appear to indicate that visit is attempting to write to hklm registry files which should notbe happening that s the only compatibility issues that cropped up the registry write does not appear in visit s source code i ran visit through a a tool that generates callstack information to discover the source of the registry write operations andit appears to be happening down in gl calls wglswapmultiplebuffers if this is truly the case i m not sure what we can doto mitigate this issue on vista binaries built with visual studio do not have the same issue also binaries built on vista using visual studio do not have this issue starting with visit we will distribute binaries built with visual studio | 1 |
815,678 | 30,567,189,246 | IssuesEvent | 2023-07-20 18:44:21 | DiscoTrayStudios/hendrix-today | https://api.github.com/repos/DiscoTrayStudios/hendrix-today | closed | Disco Tray Logo and credits | category: ui priority: high size: small type: enhancement | There should be a mention somewhere of Disco Tray Studios with our logo and a link to the website, along with the names of the student developers who worked on the app. | 1.0 | Disco Tray Logo and credits - There should be a mention somewhere of Disco Tray Studios with our logo and a link to the website, along with the names of the student developers who worked on the app. | priority | disco tray logo and credits there should be a mention somewhere of disco tray studios with our logo and a link to the website along with the names of the student developers who worked on the app | 1 |
711,897 | 24,479,110,831 | IssuesEvent | 2022-10-08 15:27:09 | JasonBock/Rocks | https://api.github.com/repos/JasonBock/Rocks | closed | Duplicating Overrides From `new` Methods | bug High Priority | ```csharp
#nullable enable
public class PropertyBuilder
{
public virtual PropertyBuilder HasAnnotation(string annotation, object? value) => default!;
}
public class PropertyBuilder<TProperty> : PropertyBuilder
{
public new virtual PropertyBuilder<TProperty> HasAnnotation(string annotation, object? value) => default!;
}
public static class Test
{
public static void Go()
{
var expectations = Rock.Create<PropertyBuilder<object>>();
}
}
```
What the mock type should generate is just the one method from the generic class, not both of them. So I'm guessing I need logic to see if a method is "shadowing" a base one via `new`. | 1.0 | Duplicating Overrides From `new` Methods - ```csharp
#nullable enable
public class PropertyBuilder
{
public virtual PropertyBuilder HasAnnotation(string annotation, object? value) => default!;
}
public class PropertyBuilder<TProperty> : PropertyBuilder
{
public new virtual PropertyBuilder<TProperty> HasAnnotation(string annotation, object? value) => default!;
}
public static class Test
{
public static void Go()
{
var expectations = Rock.Create<PropertyBuilder<object>>();
}
}
```
What the mock type should generate is just the one method from the generic class, not both of them. So I'm guessing I need logic to see if a method is "shadowing" a base one via `new`. | priority | duplicating overrides from new methods csharp nullable enable public class propertybuilder public virtual propertybuilder hasannotation string annotation object value default public class propertybuilder propertybuilder public new virtual propertybuilder hasannotation string annotation object value default public static class test public static void go var expectations rock create what the mock type should generate is just the one method from the generic class not both of them so i m guessing i need logic to see if a method is shadowing a base one via new | 1 |
707,659 | 24,313,213,472 | IssuesEvent | 2022-09-30 02:02:31 | AY2223S1-CS2103T-T15-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-T15-1/tp | closed | Update About Us Document | priority.high type.about-us | The "About Us" document/page should contain details about the team working on the project (that's us), and we should reflect that.
The document should contain a simple profile for each of us, which should include:
- A profile picture (pick a nice one)
- Your name / github username
- Your roles & responsibilities (as we have discussed)
- Anything else you want to add for fun
With this, the subtasks involved are for each person to update and add in their own details:
- Chao Yung #19
- Kevin #10
- Prittam #20
- Shawn #21
- Shi Jun #22 | 1.0 | Update About Us Document - The "About Us" document/page should contain details about the team working on the project (that's us), and we should reflect that.
The document should contain a simple profile for each of us, which should include:
- A profile picture (pick a nice one)
- Your name / github username
- Your roles & responsibilities (as we have discussed)
- Anything else you want to add for fun
With this, the subtasks involved are for each person to update and add in their own details:
- Chao Yung #19
- Kevin #10
- Prittam #20
- Shawn #21
- Shi Jun #22 | priority | update about us document the about us document page should contain details about the team working on the project that s us and we should reflect that the document should contain a simple profile for each of us which should include a profile picture pick a nice one your name github username your roles responsibilities as we have discussed anything else you want to add for fun with this the subtasks involved are for each person to update and add in their own details chao yung kevin prittam shawn shi jun | 1 |
419,147 | 12,218,234,580 | IssuesEvent | 2020-05-01 18:53:58 | nextstrain/auspice | https://api.github.com/repos/nextstrain/auspice | opened | Provide toggle to display transmission lines in the map | enhancement high priority moderate problem please take this issue | **Context**
With SARS-CoV-2 jumping around the globe so quickly, showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding. See for example:
<img width="1134" alt="lines" src="https://user-images.githubusercontent.com/1176109/80832435-807ea380-8ba1-11ea-9418-b0972f8c21f3.png">
There are frequency situations where we want "traits" reconstruction on phylogeny, but don't want to show these lines on the map.
**Description**
Include a toggle in the sidebar under "Map Options" directly below "Geographic Resolution" that says "Show transmission lines". When toggled on, we'd have the current behavior. When toggled off, the deme circles / pie charts would remain but no transmission lines would be drawn on the map.
In addition, this toggle should update the URL and (eventually) this "show transmission lines" default behavior should be included in the Auspice JSON as something like:
```
"display_defaults": {
"transmission_lines": true
},
```
But would be good to get working as UI element and URL before bothering to include in JSON. | 1.0 | Provide toggle to display transmission lines in the map - **Context**
With SARS-CoV-2 jumping around the globe so quickly, showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding. See for example:
<img width="1134" alt="lines" src="https://user-images.githubusercontent.com/1176109/80832435-807ea380-8ba1-11ea-9418-b0972f8c21f3.png">
There are frequency situations where we want "traits" reconstruction on phylogeny, but don't want to show these lines on the map.
**Description**
Include a toggle in the sidebar under "Map Options" directly below "Geographic Resolution" that says "Show transmission lines". When toggled on, we'd have the current behavior. When toggled off, the deme circles / pie charts would remain but no transmission lines would be drawn on the map.
In addition, this toggle should update the URL and (eventually) this "show transmission lines" default behavior should be included in the Auspice JSON as something like:
```
"display_defaults": {
"transmission_lines": true
},
```
But would be good to get working as UI element and URL before bothering to include in JSON. | priority | provide toggle to display transmission lines in the map context with sars cov jumping around the globe so quickly showing transmission lines on the map can be overwhelming in terms of information content and detract from overall understanding see for example img width alt lines src there are frequency situations where we want traits reconstruction on phylogeny but don t want to show these lines on the map description include a toggle in the sidebar under map options directly below geographic resolution that says show transmission lines when toggled on we d have the current behavior when toggled off the deme circles pie charts would remain but no transmission lines would be drawn on the map in addition this toggle should update the url and eventually this show transmission lines default behavior should be included in the auspice json as something like display defaults transmission lines true but would be good to get working as ui element and url before bothering to include in json | 1 |
236,838 | 7,753,052,868 | IssuesEvent | 2018-05-30 22:33:05 | Gloirin/m2gTest | https://api.github.com/repos/Gloirin/m2gTest | closed | 0004294:
renaming a resource crashes | Calendar bug high priority | **Reported by pschuele on 13 May 2011 09:24**
**Version:** Maischa (2011-05) Beta 1
renaming a resource crashes
**Additional information:** $_record is of invalid model type
.../Calendar/Controller/Resource.php(122): Tinebase_Backend_Sql_Abstract->update()
[internal function]: Calendar_Controller_Resource->update()
.../Tinebase/Frontend/Json/Abstract.php(163): call_user_func_array()
.../Calendar/Frontend/Json.php(200): Tinebase_Frontend_Json_Abstract->_save()
[internal function]: Calendar_Frontend_Json->saveResource()
.../library/Zend/Server/Abstract.php(232): call_user_func_array()
.../Zend/Json/Server.php(558): Zend_Server_Abstract->_dispatch()
.../Zend/Json/Server.php(197): Zend_Json_Server->_handle()
.../Tinebase/Server/Json.php(140): Zend_Json_Server->handle()
.../Tinebase/Server/Json.php(76): Tinebase_Server_Json->_handle()
.../Tinebase/Core.php(223): Tinebase_Server_Json->handle()
.../index.php(57): Tinebase_Core::dispatchRequest()
| 1.0 | 0004294:
renaming a resource crashes - **Reported by pschuele on 13 May 2011 09:24**
**Version:** Maischa (2011-05) Beta 1
renaming a resource crashes
**Additional information:** $_record is of invalid model type
.../Calendar/Controller/Resource.php(122): Tinebase_Backend_Sql_Abstract->update()
[internal function]: Calendar_Controller_Resource->update()
.../Tinebase/Frontend/Json/Abstract.php(163): call_user_func_array()
.../Calendar/Frontend/Json.php(200): Tinebase_Frontend_Json_Abstract->_save()
[internal function]: Calendar_Frontend_Json->saveResource()
.../library/Zend/Server/Abstract.php(232): call_user_func_array()
.../Zend/Json/Server.php(558): Zend_Server_Abstract->_dispatch()
.../Zend/Json/Server.php(197): Zend_Json_Server->_handle()
.../Tinebase/Server/Json.php(140): Zend_Json_Server->handle()
.../Tinebase/Server/Json.php(76): Tinebase_Server_Json->_handle()
.../Tinebase/Core.php(223): Tinebase_Server_Json->handle()
.../index.php(57): Tinebase_Core::dispatchRequest()
| priority | renaming a resource crashes reported by pschuele on may version maischa beta renaming a resource crashes additional information record is of invalid model type calendar controller resource php tinebase backend sql abstract gt update calendar controller resource gt update tinebase frontend json abstract php call user func array calendar frontend json php tinebase frontend json abstract gt save calendar frontend json gt saveresource library zend server abstract php call user func array zend json server php zend server abstract gt dispatch zend json server php zend json server gt handle tinebase server json php zend json server gt handle tinebase server json php tinebase server json gt handle tinebase core php tinebase server json gt handle index php tinebase core dispatchrequest | 1 |
31,120 | 2,732,045,414 | IssuesEvent | 2015-04-17 00:52:05 | EFForg/privacybadgerchrome | https://api.github.com/repos/EFForg/privacybadgerchrome | opened | Memory leak | bug High priority | We have a memory leak, in my browser privacy badger is using about 1.5G of memory. | 1.0 | Memory leak - We have a memory leak, in my browser privacy badger is using about 1.5G of memory. | priority | memory leak we have a memory leak in my browser privacy badger is using about of memory | 1 |
348,804 | 10,452,991,241 | IssuesEvent | 2019-09-19 15:40:29 | openshift/odo | https://api.github.com/repos/openshift/odo | closed | `odo component list -o json` returns empty list for `not pushed` components | kind/bug priority/High | [kind/bug]
`odo component list` should display even components that are not pushed.
The same problem is with `odo component describe` It should show information even for components that are not pushed to the cluster.
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the Google group if you have a question rather than a bug or feature request.
The group is at: https://groups.google.com/forum/#!forum/odo-users
Thanks for understanding, and for contributing to the project!
-->
## How did you run odo exactly?
`odo component list -o json | jq .`
## Actual behavior
```
▶ odo component list -o json | jq .
{
"kind": "List",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {},
"items": []
}
```
```
▶ odo component list
✗ There are no components deployed.
```
## Expected behavior
```
▶ odo component list -o json | jq .
{
"kind": "List",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {},
"items": [
{
"kind": "Component",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {
"name": "java-spring-boot-addx",
"namespace": "asdf",
"creationTimestamp": null
},
"spec": {
"app": "app",
"type": "java",
"source": "file://./",
"url": [
"java-spring-boot-addx-8080"
]
},
"status": {
"state": "Not pushed"
}
}
]
}
```
```
▶ odo component list
APP NAME TYPE SOURCE STATE
app java-spring-boot-addx java file://./ Not pushed
```
| 1.0 | `odo component list -o json` returns empty list for `not pushed` components - [kind/bug]
`odo component list` should display even components that are not pushed.
The same problem is with `odo component describe` It should show information even for components that are not pushed to the cluster.
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the Google group if you have a question rather than a bug or feature request.
The group is at: https://groups.google.com/forum/#!forum/odo-users
Thanks for understanding, and for contributing to the project!
-->
## How did you run odo exactly?
`odo component list -o json | jq .`
## Actual behavior
```
▶ odo component list -o json | jq .
{
"kind": "List",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {},
"items": []
}
```
```
▶ odo component list
✗ There are no components deployed.
```
## Expected behavior
```
▶ odo component list -o json | jq .
{
"kind": "List",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {},
"items": [
{
"kind": "Component",
"apiVersion": "odo.openshift.io/v1alpha1",
"metadata": {
"name": "java-spring-boot-addx",
"namespace": "asdf",
"creationTimestamp": null
},
"spec": {
"app": "app",
"type": "java",
"source": "file://./",
"url": [
"java-spring-boot-addx-8080"
]
},
"status": {
"state": "Not pushed"
}
}
]
}
```
```
▶ odo component list
APP NAME TYPE SOURCE STATE
app java-spring-boot-addx java file://./ Not pushed
```
| priority | odo component list o json returns empty list for not pushed components odo component list should display even components that are not pushed the same problem is with odo component describe it should show information even for components that are not pushed to the cluster welcome we kindly ask you to fill out the issue template below use the google group if you have a question rather than a bug or feature request the group is at thanks for understanding and for contributing to the project how did you run odo exactly odo component list o json jq actual behavior ▶ odo component list o json jq kind list apiversion odo openshift io metadata items ▶ odo component list ✗ there are no components deployed expected behavior ▶ odo component list o json jq kind list apiversion odo openshift io metadata items kind component apiversion odo openshift io metadata name java spring boot addx namespace asdf creationtimestamp null spec app app type java source file url java spring boot addx status state not pushed ▶ odo component list app name type source state app java spring boot addx java file not pushed | 1 |
685,779 | 23,467,034,792 | IssuesEvent | 2022-08-16 17:47:40 | larsiusprime/tdrpg-bugs | https://api.github.com/repos/larsiusprime/tdrpg-bugs | closed | The Markos/Ketta Bad Ending still plays 1 second of the sad song, then loads the normal ending music. | bug DQ CORE Cutscene 1 Please Verify Priority HIGH | I thought this had been fixed a few times in DQold. How does this keep popping up? :P
| 1.0 | The Markos/Ketta Bad Ending still plays 1 second of the sad song, then loads the normal ending music. - I thought this had been fixed a few times in DQold. How does this keep popping up? :P
| priority | the markos ketta bad ending still plays second of the sad song then loads the normal ending music i thought this had been fixed a few times in dqold how does this keep popping up p | 1 |
693,553 | 23,781,237,203 | IssuesEvent | 2022-09-02 05:12:25 | younginnovations/iatipublisher | https://api.github.com/repos/younginnovations/iatipublisher | closed | Bug :Activity Detail>>Publishing issue | type: bug priority: high Backend | Context
- Desktop
- Chrome 102.0.5005.61
Precondition
- https://stage.iatipublisher.yipl.com.np/
- Username: Publisher 3
- Password: test1234
- for created activity
- [x] **Issue 1 :After publishing and editing activity data, the user cannot unpublish the activity data.**
steps
1. publish activity data
2. After publishing activity data edit the activity data
Actual Result

Excepted Result

- [x] **Issue 2 : Error message is displayed after unpublishing the data.**
Actual Result

Excepted Result
- Error message should not displayed after unpublishing the data | 1.0 | Bug :Activity Detail>>Publishing issue - Context
- Desktop
- Chrome 102.0.5005.61
Precondition
- https://stage.iatipublisher.yipl.com.np/
- Username: Publisher 3
- Password: test1234
- for created activity
- [x] **Issue 1 :After publishing and editing activity data, the user cannot unpublish the activity data.**
steps
1. publish activity data
2. After publishing activity data edit the activity data
Actual Result

Excepted Result

- [x] **Issue 2 : Error message is displayed after unpublishing the data.**
Actual Result

Excepted Result
- Error message should not displayed after unpublishing the data | priority | bug activity detail publishing issue context desktop chrome precondition username publisher password for created activity issue after publishing and editing activity data the user cannot unpublish the activity data steps publish activity data after publishing activity data edit the activity data actual result excepted result issue error message is displayed after unpublishing the data actual result excepted result error message should not displayed after unpublishing the data | 1 |
806,863 | 29,923,872,473 | IssuesEvent | 2023-06-22 02:28:57 | TEAM-cafe-in/cafe-in-be | https://api.github.com/repos/TEAM-cafe-in/cafe-in-be | closed | feature: 카페 상세정보 조회 API를 구현한다 | 🔥High priority | ### As-is
---
-
-
### To-be
- [x] MemberInfoController response에 ViewedCafe 리스트를 반환하도록 변경한다
- [x] 커피콩 차감 로직을 구현한다. 카페 조회 시 조회한 카페 목록을 추가한다
- [x] CafeInfo API를 구현한다 | 1.0 | feature: 카페 상세정보 조회 API를 구현한다 - ### As-is
---
-
-
### To-be
- [x] MemberInfoController response에 ViewedCafe 리스트를 반환하도록 변경한다
- [x] 커피콩 차감 로직을 구현한다. 카페 조회 시 조회한 카페 목록을 추가한다
- [x] CafeInfo API를 구현한다 | priority | feature 카페 상세정보 조회 api를 구현한다 as is to be memberinfocontroller response에 viewedcafe 리스트를 반환하도록 변경한다 커피콩 차감 로직을 구현한다 카페 조회 시 조회한 카페 목록을 추가한다 cafeinfo api를 구현한다 | 1 |
176,795 | 6,565,305,835 | IssuesEvent | 2017-09-08 07:44:30 | OpenEMS/openems | https://api.github.com/repos/OpenEMS/openems | closed | refactor monitoring | Component: UI Priority: High Type: Enhancement | <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x ] feature request
```
### Bug description or desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
The Online-Monitoring should run more stable and faster | 1.0 | refactor monitoring - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x ] feature request
```
### Bug description or desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
The Online-Monitoring should run more stable and faster | priority | refactor monitoring if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request bug description or desired functionality what would like to see implemented what is the usecase the online monitoring should run more stable and faster | 1 |
39,577 | 2,856,985,822 | IssuesEvent | 2015-06-02 17:30:16 | ChrisMahlke/contribute-dev | https://api.github.com/repos/ChrisMahlke/contribute-dev | closed | Curator: once an item is in Nomcur, if it goes unshared, warn the curator | HighPriority | Example:
Rupert nominated his storymap. Deane then accepted it on production.
Rupert made a change, such that the storymap's item is no longer available from
http://www.arcgis.com/home/item.html?id=9c1100e474e44c55bf9edbe49e35fa19
The record is still visible in Nomcur table, but the app does not warn the curator.
This is low-impact, since the user knows he unshared it, but to the curator, it's as if this item never existed.
A possible solution: have an option to view records like this as a filter.
| 1.0 | Curator: once an item is in Nomcur, if it goes unshared, warn the curator - Example:
Rupert nominated his storymap. Deane then accepted it on production.
Rupert made a change, such that the storymap's item is no longer available from
http://www.arcgis.com/home/item.html?id=9c1100e474e44c55bf9edbe49e35fa19
The record is still visible in Nomcur table, but the app does not warn the curator.
This is low-impact, since the user knows he unshared it, but to the curator, it's as if this item never existed.
A possible solution: have an option to view records like this as a filter.
| priority | curator once an item is in nomcur if it goes unshared warn the curator example rupert nominated his storymap deane then accepted it on production rupert made a change such that the storymap s item is no longer available from the record is still visible in nomcur table but the app does not warn the curator this is low impact since the user knows he unshared it but to the curator it s as if this item never existed a possible solution have an option to view records like this as a filter | 1 |
335,144 | 10,149,370,454 | IssuesEvent | 2019-08-05 15:04:55 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Investigate Accuracy regression on MobileNetV2 | high priority module: vision triaged | Duplicate of https://github.com/pytorch/vision/issues/1172, but in main torch repo so we can keep all milestone issues together. | 1.0 | Investigate Accuracy regression on MobileNetV2 - Duplicate of https://github.com/pytorch/vision/issues/1172, but in main torch repo so we can keep all milestone issues together. | priority | investigate accuracy regression on duplicate of but in main torch repo so we can keep all milestone issues together | 1 |
52,205 | 3,022,259,491 | IssuesEvent | 2015-07-31 19:14:41 | EFForg/privacybadgerfirefox | https://api.github.com/repos/EFForg/privacybadgerfirefox | closed | Top level domain user settings override subdomain user settings | bug high-priority | If I set an individual subdomain to red and then the top level domain to yellow, I see that the individual subdomain gets put under the etld as cookie blocked, it should be blocked. Similarly if I have a top level domain that is set to green, any subdomains which I previously had set to yellow or red get put under the green top level domain.
So it seems like etld/green trumps subdomain red and yellow, etld yellow trumps subdomain red and etld red trumps no subdomain.
I feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it's top level domain. (An alternative argument could be made that it should always be grouped under it's parent domain, either way it should be consistent.) | 1.0 | Top level domain user settings override subdomain user settings - If I set an individual subdomain to red and then the top level domain to yellow, I see that the individual subdomain gets put under the etld as cookie blocked, it should be blocked. Similarly if I have a top level domain that is set to green, any subdomains which I previously had set to yellow or red get put under the green top level domain.
So it seems like etld/green trumps subdomain red and yellow, etld yellow trumps subdomain red and etld red trumps no subdomain.
I feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it's top level domain. (An alternative argument could be made that it should always be grouped under it's parent domain, either way it should be consistent.) | priority | top level domain user settings override subdomain user settings if i set an individual subdomain to red and then the top level domain to yellow i see that the individual subdomain gets put under the etld as cookie blocked it should be blocked similarly if i have a top level domain that is set to green any subdomains which i previously had set to yellow or red get put under the green top level domain so it seems like etld green trumps subdomain red and yellow etld yellow trumps subdomain red and etld red trumps no subdomain i feel like the correct behavior here is that if a subdomain is already in a user list it should not be grouped under it s top level domain an alternative argument could be made that it should always be grouped under it s parent domain either way it should be consistent | 1 |
67,832 | 3,282,071,787 | IssuesEvent | 2015-10-28 02:46:24 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Clicking the preference button on a filtered deck with "cards selected by relative overdueness" crashes Anki | bug fixedindev Priority-High | Originally reported on Google Code with ID 1967
```
I create a filtered deck via the desktop client that uses the relative overdueness option.
If I click on the preference button inside Ankidroid while opening that deck Ankidroid
crashes.
```
Reported by `christian.kleineidam` on 2014-01-15 14:58:46
| 1.0 | Clicking the preference button on a filtered deck with "cards selected by relative overdueness" crashes Anki - Originally reported on Google Code with ID 1967
```
I create a filtered deck via the desktop client that uses the relative overdueness option.
If I click on the preference button inside Ankidroid while opening that deck Ankidroid
crashes.
```
Reported by `christian.kleineidam` on 2014-01-15 14:58:46
| priority | clicking the preference button on a filtered deck with cards selected by relative overdueness crashes anki originally reported on google code with id i create a filtered deck via the desktop client that uses the relative overdueness option if i click on the preference button inside ankidroid while opening that deck ankidroid crashes reported by christian kleineidam on | 1 |
785,362 | 27,610,648,899 | IssuesEvent | 2023-03-09 15:46:19 | woocommerce/woocommerce-blocks | https://api.github.com/repos/woocommerce/woocommerce-blocks | opened | Third party tax calculation not displaying after changing shipping method | type: bug priority: high | ## Describe the bug
When using a third-party tax calculation plugin (AvaTax), they are not displayed on Cart and Checkout blocks after changing shipping methods. A page refresh will properly display them.
## To reproduce
Steps to reproduce the behavior:
1. Install Avatax and enable taxes
2. Create several shipping methods
3. Add items to the cart
4. Verify taxes display on the cart and checkout blocks but disappear after changing shipping methods.
## Expected behavior
Taxes should be recalculated and displayed after changing shipping methods. Checking out should include taxes.
## Screen recording
https://user-images.githubusercontent.com/17236129/224076426-cd37942f-2d3f-47b9-81df-aaade2a3a817.mov
## Additional context
Since on-page refresh and on the legacy cart and checkout taxes are correct, we need to investigate if this is a problem that needs fixing on our side.
| 1.0 | Third party tax calculation not displaying after changing shipping method - ## Describe the bug
When using a third-party tax calculation plugin (AvaTax), they are not displayed on Cart and Checkout blocks after changing shipping methods. A page refresh will properly display them.
## To reproduce
Steps to reproduce the behavior:
1. Install Avatax and enable taxes
2. Create several shipping methods
3. Add items to the cart
4. Verify taxes display on the cart and checkout blocks but disappear after changing shipping methods.
## Expected behavior
Taxes should be recalculated and displayed after changing shipping methods. Checking out should include taxes.
## Screen recording
https://user-images.githubusercontent.com/17236129/224076426-cd37942f-2d3f-47b9-81df-aaade2a3a817.mov
## Additional context
Since on-page refresh and on the legacy cart and checkout taxes are correct, we need to investigate if this is a problem that needs fixing on our side.
| priority | third party tax calculation not displaying after changing shipping method describe the bug when using a third party tax calculation plugin avatax they are not displayed on cart and checkout blocks after changing shipping methods a page refresh will properly display them to reproduce steps to reproduce the behavior install avatax and enable taxes create several shipping methods add items to the cart verify taxes display on the cart and checkout blocks but disappear after changing shipping methods expected behavior taxes should be recalculated and displayed after changing shipping methods checking out should include taxes screen recording additional context since on page refresh and on the legacy cart and checkout taxes are correct we need to investigate if this is a problem that needs fixing on our side | 1 |
295,392 | 9,085,801,102 | IssuesEvent | 2019-02-18 09:20:50 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | Assertion failure cancelling live data operation using Kafka listener | Component: Live Data Priority: High | ### Expected behavior
A live data operation should be able to be cancelled cleanly.
### Actual behavior
While testing #24799 it was observed that cancelling `MonitorLiveData` fires a debug assertion regarding an empty shared_ptr. Terminal log:
```
ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument
ConfigService-[Information] This is Mantid version 3.13.20190218.859 revision g1b57e42
ConfigService-[Information] running on ndw1213 starting 2019-02-18T09:13Z
ConfigService-[Information] Properties file(s) loaded: /media/data1/source/github/mantidproject/mantid-builds/nightly-py2/bin/Mantid.properties, /etc/mantid.local.properties, /home/dmn58364/.mantid/Mantid.user.properties
ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument
FrameworkManager-[Notice] Welcome to Mantid 3.13.20190218.859
FrameworkManager-[Notice] Please cite: http://dx.doi.org/10.1016/j.nima.2014.07.029 and this release: http://dx.doi.org/10.5286/Software/Mantid
StartLiveData-[Notice] StartLiveData started
LoadLiveData-[Notice] Performing the Replace operation.
MonitorLiveData-[Notice] MonitorLiveData started
StartLiveData-[Notice] StartLiveData successful, Duration 0.20 seconds
python: /usr/include/boost/smart_ptr/shared_ptr.hpp:648: typename boost::detail::sp_member_access<T>::type boost::shared_ptr<T>::operator->() const [with T = Mantid::DataObjects::EventWorkspace; typename boost::detail::sp_member_access<T>::type = Mantid::DataObjects::EventWorkspace*]: Assertion `px != 0' failed.
```
### Steps to reproduce the behavior
* start MantidPlot
* set facility to ISIS
* load live data
* select instrument=MARI
* select connection=kafka_event
* click run
* wait until `MonitorLiveData` has started and click cancel => :boom:
### Platforms affected
Verified on Ubuntu 16.04 but I would assume all.
| 1.0 | Assertion failure cancelling live data operation using Kafka listener - ### Expected behavior
A live data operation should be able to be cancelled cleanly.
### Actual behavior
While testing #24799 it was observed that cancelling `MonitorLiveData` fires a debug assertion regarding an empty shared_ptr. Terminal log:
```
ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument
ConfigService-[Information] This is Mantid version 3.13.20190218.859 revision g1b57e42
ConfigService-[Information] running on ndw1213 starting 2019-02-18T09:13Z
ConfigService-[Information] Properties file(s) loaded: /media/data1/source/github/mantidproject/mantid-builds/nightly-py2/bin/Mantid.properties, /etc/mantid.local.properties, /home/dmn58364/.mantid/Mantid.user.properties
ConfigService-[Information] Unable to locate directory at: /etc/mantid/instrument
FrameworkManager-[Notice] Welcome to Mantid 3.13.20190218.859
FrameworkManager-[Notice] Please cite: http://dx.doi.org/10.1016/j.nima.2014.07.029 and this release: http://dx.doi.org/10.5286/Software/Mantid
StartLiveData-[Notice] StartLiveData started
LoadLiveData-[Notice] Performing the Replace operation.
MonitorLiveData-[Notice] MonitorLiveData started
StartLiveData-[Notice] StartLiveData successful, Duration 0.20 seconds
python: /usr/include/boost/smart_ptr/shared_ptr.hpp:648: typename boost::detail::sp_member_access<T>::type boost::shared_ptr<T>::operator->() const [with T = Mantid::DataObjects::EventWorkspace; typename boost::detail::sp_member_access<T>::type = Mantid::DataObjects::EventWorkspace*]: Assertion `px != 0' failed.
```
### Steps to reproduce the behavior
* start MantidPlot
* set facility to ISIS
* load live data
* select instrument=MARI
* select connection=kafka_event
* click run
* wait until `MonitorLiveData` has started and click cancel => :boom:
### Platforms affected
Verified on Ubuntu 16.04 but I would assume all.
| priority | assertion failure cancelling live data operation using kafka listener expected behavior a live data operation should be able to be cancelled cleanly actual behavior while testing it was observed that cancelling monitorlivedata fires a debug assertion regarding an empty shared ptr terminal log configservice unable to locate directory at etc mantid instrument configservice this is mantid version revision configservice running on starting configservice properties file s loaded media source github mantidproject mantid builds nightly bin mantid properties etc mantid local properties home mantid mantid user properties configservice unable to locate directory at etc mantid instrument frameworkmanager welcome to mantid frameworkmanager please cite and this release startlivedata startlivedata started loadlivedata performing the replace operation monitorlivedata monitorlivedata started startlivedata startlivedata successful duration seconds python usr include boost smart ptr shared ptr hpp typename boost detail sp member access type boost shared ptr operator const assertion px failed steps to reproduce the behavior start mantidplot set facility to isis load live data select instrument mari select connection kafka event click run wait until monitorlivedata has started and click cancel boom platforms affected verified on ubuntu but i would assume all | 1 |
434,477 | 12,518,897,877 | IssuesEvent | 2020-06-03 13:40:04 | carbon-design-system/ibm-dotcom-library | https://api.github.com/repos/carbon-design-system/ibm-dotcom-library | closed | Create codesandbox for Dotcom Shell (Vanilla) | Airtable Done dev package: vanilla priority: high | <!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
IBM.com Library adopter
> I need to:
see the DotcomShell in an isolated codesandbox
> so that I can:
see how I can easily integrate into my application
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- Should create under the `ibmdotcom-bot` github user under https://codesandbox.io
- 1Password access for `ibmdotcom-bot`: https://ibm.ent.box.com/notes/496014563545
- Template available under codesandbox: https://codesandbox.io/dashboard/templates
#### Acceptance criteria
- [ ] Codesandbox created for DotcomShell
- [ ] Link added to README
| 1.0 | Create codesandbox for Dotcom Shell (Vanilla) - <!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
IBM.com Library adopter
> I need to:
see the DotcomShell in an isolated codesandbox
> so that I can:
see how I can easily integrate into my application
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- Should create under the `ibmdotcom-bot` github user under https://codesandbox.io
- 1Password access for `ibmdotcom-bot`: https://ibm.ent.box.com/notes/496014563545
- Template available under codesandbox: https://codesandbox.io/dashboard/templates
#### Acceptance criteria
- [ ] Codesandbox created for DotcomShell
- [ ] Link added to README
| priority | create codesandbox for dotcom shell vanilla user story as a ibm com library adopter i need to see the dotcomshell in an isolated codesandbox so that i can see how i can easily integrate into my application additional information should create under the ibmdotcom bot github user under access for ibmdotcom bot template available under codesandbox acceptance criteria codesandbox created for dotcomshell link added to readme | 1 |
386,304 | 11,435,062,276 | IssuesEvent | 2020-02-04 18:38:09 | collinbarrett/FilterLists | https://api.github.com/repos/collinbarrett/FilterLists | closed | FilterLists Hub isn't working | bug high priority | I see only dark background on FilterLists Hub and errors on console. | 1.0 | FilterLists Hub isn't working - I see only dark background on FilterLists Hub and errors on console. | priority | filterlists hub isn t working i see only dark background on filterlists hub and errors on console | 1 |
663,090 | 22,161,700,081 | IssuesEvent | 2022-06-04 15:46:11 | ApplETS/Notre-Dame | https://api.github.com/repos/ApplETS/Notre-Dame | opened | Rating does not work on button click | bug platform: ios platform: android ready to develop priority: high | **Describe the bug**
The button 'rate me' don't do anything.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'More'
2. Click on 'Rate Me'
3. See error
**Expected behavior**
The button should show the in app review or at least redirect to store listing.
**Screenshots**
N/A
**Additional context**
N/A
| 1.0 | Rating does not work on button click - **Describe the bug**
The button 'rate me' don't do anything.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'More'
2. Click on 'Rate Me'
3. See error
**Expected behavior**
The button should show the in app review or at least redirect to store listing.
**Screenshots**
N/A
**Additional context**
N/A
| priority | rating does not work on button click describe the bug the button rate me don t do anything to reproduce steps to reproduce the behavior go to more click on rate me see error expected behavior the button should show the in app review or at least redirect to store listing screenshots n a additional context n a | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.